Date: (Mon) Jun 13, 2016

Introduction:

Data: Source: Training: “https://inclass.kaggle.com/c/can-we-predict-voting-outcomes/download/train2016.csv
New: “https://inclass.kaggle.com/c/can-we-predict-voting-outcomes/download/test2016.csv
Time period:

Synopsis:

Based on analysis utilizing <> techniques, :

Summary of key steps & error improvement stats:

Prediction Accuracy Enhancement Options:

  • transform.data chunk:
    • derive features from multiple features
  • manage.missing.data chunk:
    • Not fill missing vars
    • Fill missing numerics with a different algorithm
    • Fill missing chars with data based on clusters

[](.png)

Potential next steps include:

  • Organization:
    • Categorize by chunk
    • Priority criteria:
      1. Ease of change
      2. Impacts report
      3. Cleans innards
      4. Bug report
  • all chunks:
    • at chunk-end rm(!glb_)
  • manage.missing.data chunk:
    • cleaner way to manage re-splitting of training vs. new entity
  • extract.features chunk:
    • Add n-grams for glbFeatsText
      • “RTextTools”, “tau”, “RWeka”, and “textcat” packages
  • fit.models chunk:
    • Classification: Plot AUC Curves for all models & highlight glbMdlSel
    • Prediction accuracy scatter graph:
    • Add tiles (raw vs. PCA)
    • Use shiny for drop-down of “important” features
    • Use plot.ly for interactive plots ?

    • Change .fit suffix of model metrics to .mdl if it’s data independent (e.g. AIC, Adj.R.Squared - is it truly data independent ?, etc.)
    • create a custom model for rpart that has minbucket as a tuning parameter
    • varImp for randomForest crashes in caret version:6.0.41 -> submit bug report

  • Probability handling for multinomials vs. desired binomial outcome
  • ROCR currently supports only evaluation of binary classification tasks (version 1.0.7)
  • extensions toward multiclass classification are scheduled for the next release

  • fit.all.training chunk:
    • myplot_prediction_classification: displays ‘x’ instead of ‘+’ when there are no prediction errors
  • Compare glb_sel_mdl vs. glb_fin_mdl:
    • varImp
    • Prediction differences (shd be minimal ?)
  • Move glb_analytics_diag_plots to mydsutils.R: (+) Easier to debug (-) Too many glb vars used
  • Add print(ggplot.petrinet(glb_analytics_pn) + coord_flip()) at the end of every major chunk
  • Parameterize glb_analytics_pn
  • Move glb_impute_missing_data to mydsutils.R: (-) Too many glb vars used; glb_<>_df reassigned
  • Do non-glm methods handle interaction terms ?
  • f-score computation for classifiers should be summation across outcomes (not just the desired one ?)
  • Add accuracy computation to glb_dmy_mdl in predict.data.new chunk
  • Why does splitting fit.data.training.all chunk into separate chunks add an overhead of ~30 secs ? It’s not rbind b/c other chunks have lower elapsed time. Is it the number of plots ?
  • Incorporate code chunks in print_sessionInfo
  • Test against
    • projects in github.com/bdanalytics
    • lectures in jhu-datascience track

Analysis:

rm(list = ls())
set.seed(12345)
options(stringsAsFactors = FALSE)
source("~/Dropbox/datascience/R/mycaret.R")
source("~/Dropbox/datascience/R/mypetrinet.R")
source("~/Dropbox/datascience/R/myplclust.R")
source("~/Dropbox/datascience/R/myplot.R")
source("~/Dropbox/datascience/R/myscript.R")
source("~/Dropbox/datascience/R/mytm.R")
if (is.null(knitr::opts_current$get(name = 'label'))) # Running in IDE
    debugSource("~/Dropbox/datascience/R/mydsutils.R") else
    source("~/Dropbox/datascience/R/mydsutils.R")    
## Loading required package: caret
## Loading required package: lattice
# Gather all package requirements here
suppressPackageStartupMessages(require(doMC))
glbCores <- 10 # of cores on machine - 2
registerDoMC(glbCores) 

suppressPackageStartupMessages(require(caret))
require(plyr)
## Loading required package: plyr
require(dplyr)
## Loading required package: dplyr
## 
## Attaching package: 'dplyr'
## The following objects are masked from 'package:plyr':
## 
##     arrange, count, desc, failwith, id, mutate, rename, summarise,
##     summarize
## The following objects are masked from 'package:stats':
## 
##     filter, lag
## The following objects are masked from 'package:base':
## 
##     intersect, setdiff, setequal, union
require(knitr)
## Loading required package: knitr
require(stringr)
## Loading required package: stringr
#source("dbgcaret.R")
#packageVersion("snow")
#require(sos); findFn("cosine", maxPages=2, sortby="MaxScore")

# Analysis control global variables
# Inputs
#   url/name = "<PathPointer>"; if url specifies a zip file, name = "<filename>"; 
#               or named collection of <PathPointer>s
#   sep = choose from c(NULL, "\t")
glbObsTrnFile <- list(url = "https://inclass.kaggle.com/c/can-we-predict-voting-outcomes/download/train2016.csv"
    # or list(url = c(NULL, <.inp1> = "<path1>", <.inp2> = "<path2>"))
    #, splitSpecs = list(method = "copy" # default when glbObsNewFile is NULL
    #                       select from c("copy", NULL ???, "condition", "sample", )
    #                      ,nRatio = 0.3 # > 0 && < 1 if method == "sample" 
    #                      ,seed = 123 # any integer or glbObsTrnPartitionSeed if method == "sample" 
    #                      ,condition = # or 'is.na(<var>)'; '<var> <condition_operator> <value>'    
    #                      )
    )                   
 
glbObsNewFile <- list(url = "https://inclass.kaggle.com/c/can-we-predict-voting-outcomes/download/test2016.csv") 

glbObsDropCondition <- #NULL # : default
#   enclose in single-quotes b/c condition might include double qoutes
#       use | & ; NOT || &&    
#   '<condition>' 
    # 'grepl("^First Draft Video:", glbObsAll$Headline)'
    # 'is.na(glbObsAll[, glb_rsp_var_raw])'
    # '(is.na(glbObsAll[, glb_rsp_var_raw]) & grepl("Train", glbObsAll[, glbFeatsId]))'
    # 'is.na(strptime(glbObsAll[, "Date"], glbFeatsDateTime[["Date"]]["format"], tz = glbFeatsDateTime[["Date"]]["timezone"]))'
# '(is.na(glbObsAll[, "Q109244"]) | (glbObsAll[, "Q109244"] != "No"))' # No
# '(glbObsAll[, "Q109244"] != "")' # NA
'(is.na(glbObsAll[, "Q109244"]) | (glbObsAll[, "Q109244"] != "Yes"))' # Yes
#nrow(do.call("subset",list(glbObsAll, parse(text=paste0("!(", glbObsDropCondition, ")")))))
    
glb_obs_repartition_train_condition <- NULL # : default
#    "<condition>" 

glb_max_fitobs <- NULL # or any integer
glbObsTrnPartitionSeed <- 123 # or any integer
                         
glb_is_regression <- FALSE; glb_is_classification <- !glb_is_regression; 
    glb_is_binomial <- TRUE # or TRUE or FALSE

glb_rsp_var_raw <- "Party"

# for classification, the response variable has to be a factor
glb_rsp_var <- "Party.fctr"

# if the response factor is based on numbers/logicals e.g (0/1 OR TRUE/FALSE vs. "A"/"B"), 
#   or contains spaces (e.g. "Not in Labor Force")
#   caret predict(..., type="prob") crashes
glb_map_rsp_raw_to_var <- #NULL 
function(raw) {
#     return(raw ^ 0.5)
#     return(log(raw))
#     return(log(1 + raw))
#     return(log10(raw)) 
#     return(exp(-raw / 2))
#     
# chk ref value against frequencies vs. alpha sort order
    ret_vals <- rep_len(NA, length(raw)); ret_vals[!is.na(raw)] <- ifelse(raw[!is.na(raw)] == "Republican", "R", "D"); return(relevel(as.factor(ret_vals), ref = "D")) 
    
#     as.factor(paste0("B", raw))
#     as.factor(gsub(" ", "\\.", raw))
    }

#if glb_rsp_var_raw is numeric:
#print(summary(glbObsAll[, glb_rsp_var_raw]))
#glb_map_rsp_raw_to_var(tst <- c(NA, as.numeric(summary(glbObsAll[, glb_rsp_var_raw])))) 

#if glb_rsp_var_raw is character:
#print(table(glbObsAll[, glb_rsp_var_raw], useNA = "ifany"))
# print(table(glb_map_rsp_raw_to_var(tst <- glbObsAll[, glb_rsp_var_raw]), useNA = "ifany"))

glb_map_rsp_var_to_raw <- #NULL 
function(var) {
#     return(var ^ 2.0)
#     return(exp(var))
#     return(10 ^ var) 
#     return(-log(var) * 2)
#     as.numeric(var)
#     levels(var)[as.numeric(var)]
    sapply(levels(var)[as.numeric(var)], function(elm) 
        if (is.na(elm)) return(elm) else
        if (elm == 'R') return("Republican") else
        if (elm == 'D') return("Democrat") else
        stop("glb_map_rsp_var_to_raw: unexpected value: ", elm)
        )  
#     gsub("\\.", " ", levels(var)[as.numeric(var)])
#     c("<=50K", " >50K")[as.numeric(var)]
#     c(FALSE, TRUE)[as.numeric(var)]
}
# print(table(glb_map_rsp_var_to_raw(glb_map_rsp_raw_to_var(tst)), useNA = "ifany"))

if ((glb_rsp_var != glb_rsp_var_raw) && is.null(glb_map_rsp_raw_to_var))
    stop("glb_map_rsp_raw_to_var function expected")

# List info gathered for various columns
# <col_name>:   <description>; <notes>
# USER_ID - an anonymous id unique to a given user
# YOB - the year of birth of the user
# Gender - the gender of the user, either Male or Female
# Income - the household income of the user. Either not provided, or one of "under $25,000", "$25,001 - $50,000", "$50,000 - $74,999", "$75,000 - $100,000", "$100,001 - $150,000", or "over $150,000".
# HouseholdStatus - the household status of the user. Either not provided, or one of "Domestic Partners (no kids)", "Domestic Partners (w/kids)", "Married (no kids)", "Married (w/kids)", "Single (no kids)", or "Single (w/kids)".
# EducationalLevel - the education level of the user. Either not provided, or one of "Current K-12", "High School Diploma", "Current Undergraduate", "Associate's Degree", "Bachelor's Degree", "Master's Degree", or "Doctoral Degree".
# Party - the political party for whom the user intends to vote for. Either "Democrat" or "Republican
# Q124742, Q124122, . . . , Q96024 - 101 different questions that the users were asked on Show of Hands. If the user didn't answer the question, there is a blank. For information about the question text and possible answers, see the file Questions.pdf.

# currently does not handle more than 1 column; consider concatenating multiple columns
# If glbFeatsId == NULL, ".rownames <- as.numeric(row.names())" is the default
glbFeatsId <- "USER_ID" # choose from c(NULL : default, "<id_feat>") 
glbFeatsCategory <- "Hhold.fctr" # choose from c(NULL : default, "<category_feat>")
# glbFeatsCategory <- "Q109244.fctr" # choose from c(NULL : default, "<category_feat>")
# glbFeatsCategory <- "Q115611.fctr" # choose from c(NULL : default, "<category_feat>")

# User-specified exclusions
glbFeatsExclude <- c(NULL
#   Feats that shd be excluded due to known causation by prediction variable
# , "<feat1", "<feat2>"
#   Feats that are factors with unique values (as % of nObs) > 49 (empirically derived)
#   Feats that are linear combinations (alias in glm)
#   Feature-engineering phase -> start by excluding all features except id & category & 
#       work each one in
    , "USER_ID", "YOB", "Gender", "Income", "HouseholdStatus", "EducationLevel" 
    ,"Q124742","Q124122" 
    ,"Q123621","Q123464"
    ,"Q122771","Q122770","Q122769","Q122120"
    ,"Q121700","Q121699","Q121011"
    ,"Q120978","Q120650","Q120472","Q120379","Q120194","Q120014","Q120012" 
    ,"Q119851","Q119650","Q119334"
    ,"Q118892","Q118237","Q118233","Q118232","Q118117"
    ,"Q117193","Q117186"
    ,"Q116797","Q116881","Q116953","Q116601","Q116441","Q116448","Q116197"
    ,"Q115602","Q115777","Q115610","Q115611","Q115899","Q115390","Q115195"
    ,"Q114961","Q114748","Q114517","Q114386","Q114152"
    ,"Q113992","Q113583","Q113584","Q113181"
    ,"Q112478","Q112512","Q112270"
    ,"Q111848","Q111580","Q111220"
    ,"Q110740"
    ,"Q109367","Q109244"
    ,"Q108950","Q108855","Q108617","Q108856","Q108754","Q108342","Q108343"
    ,"Q107869","Q107491"
    ,"Q106993","Q106997","Q106272","Q106388","Q106389","Q106042"
    ,"Q105840","Q105655"
    ,"Q104996"
    ,"Q103293"
    ,"Q102906","Q102674","Q102687","Q102289","Q102089"
    ,"Q101162","Q101163","Q101596"
    ,"Q100689","Q100680","Q100562","Q100010"
    ,"Q99982"
    ,"Q99716"
    ,"Q99581"
    ,"Q99480"
    ,"Q98869"
    ,"Q98578"
    ,"Q98197"
    ,"Q98059","Q98078"
    ,"Q96024" # Done
    ,".pos") 
if (glb_rsp_var_raw != glb_rsp_var)
    glbFeatsExclude <- union(glbFeatsExclude, glb_rsp_var_raw)                    

glbFeatsInteractionOnly <- list()
#glbFeatsInteractionOnly[["<child_feat>"]] <- "<parent_feat>"
glbFeatsInteractionOnly[["YOB.Age.dff"]] <- "YOB.Age.fctr"

glbFeatsDrop <- c(NULL
                # , "<feat1>", "<feat2>"
                )

glb_map_vars <- NULL # or c("<var1>", "<var2>")
glb_map_urls <- list();
# glb_map_urls[["<var1>"]] <- "<var1.url>"

# Derived features; Use this mechanism to cleanse data ??? Cons: Data duplication ???
glbFeatsDerive <- list();

# glbFeatsDerive[["<feat.my.sfx>"]] <- list(
#     mapfn = function(<arg1>, <arg2>) { return(function(<arg1>, <arg2>)) } 
#   , args = c("<arg1>", "<arg2>"))
#myprint_df(data.frame(ImageId = mapfn(glbObsAll$.src, glbObsAll$.pos)))
#data.frame(ImageId = mapfn(glbObsAll$.src, glbObsAll$.pos))[7045:7055, ]

    # character
#     mapfn = function(Education) { raw <- Education; raw[is.na(raw)] <- "NA.my"; return(as.factor(raw)) } 
#     mapfn = function(Week) { return(substr(Week, 1, 10)) }
#     mapfn = function(Name) { return(sapply(Name, function(thsName) 
#                                             str_sub(unlist(str_split(thsName, ","))[1], 1, 1))) } 

#     mapfn = function(descriptor) { return(plyr::revalue(descriptor, c(
#         "ABANDONED BUILDING"  = "OTHER",
#         "**"                  = "**"
#                                           ))) }

#     mapfn = function(description) { mod_raw <- description;
    # This is here because it does not work if it's in txt_map_filename
#         mod_raw <- gsub(paste0(c("\n", "\211", "\235", "\317", "\333"), collapse = "|"), " ", mod_raw)
    # Don't parse for "." because of ".com"; use customized gsub for that text
#         mod_raw <- gsub("(\\w)(!|\\*|,|-|/)(\\w)", "\\1\\2 \\3", mod_raw);
    # Some state acrnoyms need context for separation e.g. 
    #   LA/L.A. could either be "Louisiana" or "LosAngeles"
        # modRaw <- gsub("\\bL\\.A\\.( |,|')", "LosAngeles\\1", modRaw);
    #   OK/O.K. could either be "Oklahoma" or "Okay"
#         modRaw <- gsub("\\bACA OK\\b", "ACA OKay", modRaw); 
#         modRaw <- gsub("\\bNow O\\.K\\.\\b", "Now OKay", modRaw);        
    #   PR/P.R. could either be "PuertoRico" or "Public Relations"        
        # modRaw <- gsub("\\bP\\.R\\. Campaign", "PublicRelations Campaign", modRaw);        
    #   VA/V.A. could either be "Virginia" or "VeteransAdministration"        
        # modRaw <- gsub("\\bthe V\\.A\\.\\:", "the VeteranAffairs:", modRaw);
    #   
    # Custom mods

#         return(mod_raw) }

    # numeric
# Create feature based on record position/id in data   
glbFeatsDerive[[".pos"]] <- list(
    mapfn = function(raw1) { return(1:length(raw1)) }
    , args = c(".rnorm"))
# glbFeatsDerive[[".pos.y"]] <- list(
#     mapfn = function(raw1) { return(1:length(raw1)) }       
#     , args = c(".rnorm"))    

# Add logs of numerics that are not distributed normally
#   Derive & keep multiple transformations of the same feature, if normality is hard to achieve with just one transformation
#   Right skew: logp1; sqrt; ^ 1/3; logp1(logp1); log10; exp(-<feat>/constant)
# glbFeatsDerive[["WordCount.log1p"]] <- list(
#     mapfn = function(WordCount) { return(log1p(WordCount)) } 
#   , args = c("WordCount"))
# glbFeatsDerive[["WordCount.root2"]] <- list(
#     mapfn = function(WordCount) { return(WordCount ^ (1/2)) } 
#   , args = c("WordCount"))
# glbFeatsDerive[["WordCount.nexp"]] <- list(
#     mapfn = function(WordCount) { return(exp(-WordCount)) } 
#   , args = c("WordCount"))
#print(summary(glbObsAll$WordCount))
#print(summary(mapfn(glbObsAll$WordCount)))
    
# If imputation shd be skipped for this feature
# glbFeatsDerive[["District.fctr"]] <- list(
#     mapfn = function(District) {
#         raw <- District;
#         ret_vals <- rep_len("NA", length(raw)); 
#         ret_vals[!is.na(raw)] <- sapply(raw[!is.na(raw)], function(elm) 
#                                         ifelse(elm < 10, "1-9", 
#                                         ifelse(elm < 20, "10-19", "20+")));
#         return(relevel(as.factor(ret_vals), ref = "NA"))
#     }       
#     , args = c("District"))    

# YOB options:
# 1. Missing data:
# 1.1   0 -> Does not improve baseline
# 1.2   Cut factors & "NA" is a level
# 2. Data corrections: < 1928 & > 2000
# 3. Scale YOB
# 4. Add Age
# YOB.Age.fctr needs to be synced with YOB.Age.dff; Create a separate sub-function ???
glbFeatsDerive[["YOB.Age.fctr"]] <- list(
    mapfn = function(raw1) {
        raw <- 2016 - raw1 
        # raw[!is.na(raw) & raw >= 2010] <- NA
        raw[!is.na(raw) & (raw <= 15)] <- NA
        raw[!is.na(raw) & (raw >= 90)] <- NA        
        retVal <- rep_len("NA", length(raw))
        # breaks = c(1879, seq(1949, 1989, 10), 2049)
        # cutVal <- cut(raw[!is.na(raw)], breaks = breaks, 
        #               labels = as.character(breaks + 1)[1:(length(breaks) - 1)])
        cutVal <- cut(raw[!is.na(raw)], breaks = c(15, 20, 25, 30, 35, 40, 50, 65, 90))
        retVal[!is.na(raw)] <- levels(cutVal)[cutVal]
        return(factor(retVal, levels = c("NA"
                ,"(15,20]","(20,25]","(25,30]","(30,35]","(35,40]","(40,50]","(50,65]","(65,90]"),
                        ordered = TRUE))
    }
    , args = c("YOB"))

# YOB.Age.fctr needs to be synced with YOB.Age.dff; Create a separate sub-function ???
glbFeatsDerive[["YOB.Age.dff"]] <- list(
    mapfn = function(raw1) {
        raw <- 2016 - raw1 
        raw[!is.na(raw) & (raw <= 15)] <- NA
        raw[!is.na(raw) & (raw >= 90)] <- NA        
        breaks <- c(15, 20, 25, 30, 35, 40, 50, 65, 90)

        # retVal <- rep_len(0, length(raw))
        stopifnot(sum(!is.na(raw) && (raw <= 15)) == 0)
        stopifnot(sum(!is.na(raw) && (raw >= 90)) == 0) 
        # msk <- !is.na(raw) && (raw > 15) && (raw <= 20); if (sum(msk > 0)) retVal[msk] <- raw[msk] - 15
        # msk <- !is.na(raw) && (raw > 20) && (raw <= 25); if (sum(msk > 0)) retVal[msk] <- raw[msk] - 20
        # msk <- !is.na(raw) && (raw > 25) && (raw <= 30); if (sum(msk > 0)) retVal[msk] <- raw[msk] - 25
        # msk <- !is.na(raw) && (raw > 30) && (raw <= 35); if (sum(msk > 0)) retVal[msk] <- raw[msk] - 30
        # msk <- !is.na(raw) && (raw > 35) && (raw <= 40); if (sum(msk > 0)) retVal[msk] <- raw[msk] - 35
        # msk <- !is.na(raw) && (raw > 40) && (raw <= 50); if (sum(msk > 0)) retVal[msk] <- raw[msk] - 40
        # msk <- !is.na(raw) && (raw > 50) && (raw <= 65); if (sum(msk > 0)) retVal[msk] <- raw[msk] - 50
        # msk <- !is.na(raw) && (raw > 65) && (raw <= 90); if (sum(msk > 0)) retVal[msk] <- raw[msk] - 65

        breaks <- c(15, 20, 25, 30, 35, 40, 50, 65, 90)        
        retVal <- sapply(raw, function(age) {
            if (is.na(age)) return(0) else
            if ((age > 15) && (age <= 20)) return(age - 15) else
            if ((age > 20) && (age <= 25)) return(age - 20) else
            if ((age > 25) && (age <= 30)) return(age - 25) else
            if ((age > 30) && (age <= 35)) return(age - 30) else
            if ((age > 35) && (age <= 40)) return(age - 35) else
            if ((age > 40) && (age <= 50)) return(age - 40) else
            if ((age > 50) && (age <= 65)) return(age - 50) else
            if ((age > 65) && (age <= 90)) return(age - 65)
        })
        
        return(retVal)
    }
    , args = c("YOB"))

glbFeatsDerive[["Gender.fctr"]] <- list(
    mapfn = function(raw1) {
        raw <- raw1
        raw[raw %in% ""] <- "N"
        raw <- gsub("Male"  , "M", raw, fixed = TRUE)
        raw <- gsub("Female", "F", raw, fixed = TRUE)        
        return(relevel(as.factor(raw), ref = "N"))
    }
    , args = c("Gender"))

glbFeatsDerive[["Income.fctr"]] <- list(
    mapfn = function(raw1) { raw <- raw1;
        raw[raw %in% ""] <- "N"
        raw <- gsub("under $25,000"      , "<25K"    , raw, fixed = TRUE)
        raw <- gsub("$25,001 - $50,000"  , "25-50K"  , raw, fixed = TRUE)
        raw <- gsub("$50,000 - $74,999"  , "50-75K"  , raw, fixed = TRUE)
        raw <- gsub("$75,000 - $100,000" , "75-100K" , raw, fixed = TRUE)        
        raw <- gsub("$100,001 - $150,000", "100-150K", raw, fixed = TRUE)
        raw <- gsub("over $150,000"      , ">150K"   , raw, fixed = TRUE)        
        return(factor(raw, levels = c("N","<25K","25-50K","50-75K","75-100K","100-150K",">150K"),
                      ordered = TRUE))
    }
    , args = c("Income"))

glbFeatsDerive[["Hhold.fctr"]] <- list(
    mapfn = function(raw1) { raw <- raw1;
        raw[raw %in% ""] <- "N"
        raw <- gsub("Domestic Partners (no kids)", "PKn", raw, fixed = TRUE)
        raw <- gsub("Domestic Partners (w/kids)" , "PKy", raw, fixed = TRUE)        
        raw <- gsub("Married (no kids)"          , "MKn", raw, fixed = TRUE)
        raw <- gsub("Married (w/kids)"           , "MKy", raw, fixed = TRUE)        
        raw <- gsub("Single (no kids)"           , "SKn", raw, fixed = TRUE)
        raw <- gsub("Single (w/kids)"            , "SKy", raw, fixed = TRUE)        
        return(relevel(as.factor(raw), ref = "N"))
    }
    , args = c("HouseholdStatus"))

glbFeatsDerive[["Edn.fctr"]] <- list(
    mapfn = function(raw1) { raw <- raw1;
        raw[raw %in% ""] <- "N"
        raw <- gsub("Current K-12"         , "K12", raw, fixed = TRUE)
        raw <- gsub("High School Diploma"  , "HSD", raw, fixed = TRUE)        
        raw <- gsub("Current Undergraduate", "CCg", raw, fixed = TRUE)
        raw <- gsub("Associate's Degree"   , "Ast", raw, fixed = TRUE)
        raw <- gsub("Bachelor's Degree"    , "Bcr", raw, fixed = TRUE)        
        raw <- gsub("Master's Degree"      , "Msr", raw, fixed = TRUE)
        raw <- gsub("Doctoral Degree"      , "PhD", raw, fixed = TRUE)        
        return(factor(raw, levels = c("N","K12","HSD","CCg","Ast","Bcr","Msr","PhD"),
                      ordered = TRUE))
    }
    , args = c("EducationLevel"))

# for (qsn in c("Q124742","Q124122"))
# for (qsn in grep("Q12(.{4})(?!\\.fctr)", names(glbObsTrn), value = TRUE, perl = TRUE))
for (qsn in grep("Q", glbFeatsExclude, fixed = TRUE, value = TRUE))    
    glbFeatsDerive[[paste0(qsn, ".fctr")]] <- list(
        mapfn = function(raw1) {
            raw1[raw1 %in% ""] <- "NA"
            rawVal <- unique(raw1)
            
            if (length(setdiff(rawVal, (expVal <- c("NA", "No", "Ys")))) == 0) {
                raw1 <- gsub("Yes", "Ys", raw1, fixed = TRUE)
                if (length(setdiff(rawVal, expVal)) > 0)
                    stop(qsn, " vals: ", paste0(rawVal, collapse = "|"), 
                         " does not match expectation: ", paste0(expVal, collapse = "|"))
            } else
            if (length(setdiff(rawVal, (expVal <- c("NA", "Me", "Circumstances")))) == 0) {
                raw1 <- gsub("Circumstances", "Cs", raw1, fixed = TRUE)
                if (length(setdiff(rawVal, expVal)) > 0)
                    stop(qsn, " vals: ", paste0(rawVal, collapse = "|"), 
                         " does not match expectation: ", paste0(expVal, collapse = "|"))
            } else
            if (length(setdiff(rawVal, (expVal <- c("NA", "Grrr people", "Yay people!")))) == 0) {
                raw1 <- gsub("Grrr people", "Gr", raw1, fixed = TRUE)
                raw1 <- gsub("Yay people!", "Yy", raw1, fixed = TRUE)                
                if (length(setdiff(rawVal, expVal)) > 0)
                    stop(qsn, " vals: ", paste0(rawVal, collapse = "|"), 
                         " does not match expectation: ", paste0(expVal, collapse = "|"))
            } else
            if (length(setdiff(rawVal, (expVal <- c("NA", "Idealist", "Pragmatist")))) == 0) {
                raw1 <- gsub("Idealist"  , "Id", raw1, fixed = TRUE)
                raw1 <- gsub("Pragmatist", "Pr", raw1, fixed = TRUE)                
                if (length(setdiff(rawVal, expVal)) > 0)
                    stop(qsn, " vals: ", paste0(rawVal, collapse = "|"), 
                         " does not match expectation: ", paste0(expVal, collapse = "|"))
            } else
            if (length(setdiff(rawVal, (expVal <- c("NA", "Private", "Public")))) == 0) {
                raw1 <- gsub("Private", "Pt", raw1, fixed = TRUE)
                raw1 <- gsub("Public" , "Pc", raw1, fixed = TRUE)                
                if (length(setdiff(rawVal, expVal)) > 0)
                    stop(qsn, " vals: ", paste0(rawVal, collapse = "|"), 
                         " does not match expectation: ", paste0(expVal, collapse = "|"))
            }
            
            return(relevel(as.factor(raw1), ref = "NA"))
        }
        , args = c(qsn))

# If imputation of missing data is not working ...
# glbFeatsDerive[["FertilityRate.nonNA"]] <- list(
#     mapfn = function(FertilityRate, Region) {
#         RegionMdn <- tapply(FertilityRate, Region, FUN = median, na.rm = TRUE)
# 
#         retVal <- FertilityRate
#         retVal[is.na(FertilityRate)] <- RegionMdn[Region[is.na(FertilityRate)]]
#         return(retVal)
#     }
#     , args = c("FertilityRate", "Region"))
    
#     mapfn = function(HOSPI.COST) { return(cut(HOSPI.COST, 5, breaks = c(0, 100000, 200000, 300000, 900000), labels = NULL)) }     
#     mapfn = function(Rasmussen)  { return(ifelse(sign(Rasmussen) >= 0, 1, 0)) } 
#     mapfn = function(startprice) { return(startprice ^ (1/2)) }       
#     mapfn = function(startprice) { return(log(startprice)) }   
#     mapfn = function(startprice) { return(exp(-startprice / 20)) }
#     mapfn = function(startprice) { return(scale(log(startprice))) }     
#     mapfn = function(startprice) { return(sign(sprice.predict.diff) * (abs(sprice.predict.diff) ^ (1/10))) }        

    # factor      
#     mapfn = function(PropR) { return(as.factor(ifelse(PropR >= 0.5, "Y", "N"))) }
#     mapfn = function(productline, description) { as.factor(gsub(" ", "", productline)) }
#     mapfn = function(purpose) { return(relevel(as.factor(purpose), ref="all_other")) }
#     mapfn = function(raw) { tfr_raw <- as.character(cut(raw, 5)); 
#                             tfr_raw[is.na(tfr_raw)] <- "NA.my";
#                             return(as.factor(tfr_raw)) }
#     mapfn = function(startprice.log10) { return(cut(startprice.log10, 3)) }
#     mapfn = function(startprice.log10) { return(cut(sprice.predict.diff, c(-1000, -100, -10, -1, 0, 1, 10, 100, 1000))) }    

#     , args = c("<arg1>"))
    
    # multiple args
#     mapfn = function(id, date) { return(paste(as.character(id), as.character(date), sep = "#")) }        
#     mapfn = function(PTS, oppPTS) { return(PTS - oppPTS) }
#     mapfn = function(startprice.log10.predict, startprice) {
#                  return(spdiff <- (10 ^ startprice.log10.predict) - startprice) } 
#     mapfn = function(productline, description) { as.factor(
#         paste(gsub(" ", "", productline), as.numeric(nchar(description) > 0), sep = "*")) }
#     mapfn = function(.src, .pos) { 
#         return(paste(.src, sprintf("%04d", 
#                                    ifelse(.src == "Train", .pos, .pos - 7049)
#                                    ), sep = "#")) }       

# # If glbObsAll is not sorted in the desired manner
#     mapfn=function(Week) { return(coredata(lag(zoo(orderBy(~Week, glbObsAll)$ILI), -2, na.pad=TRUE))) }
#     mapfn=function(ILI) { return(coredata(lag(zoo(ILI), -2, na.pad=TRUE))) }
#     mapfn=function(ILI.2.lag) { return(log(ILI.2.lag)) }

# glbFeatsDerive[["<var1>"]] <- glbFeatsDerive[["<var2>"]]

# tst <- "descr.my"; args_lst <- NULL; for (arg in glbFeatsDerive[[tst]]$args) args_lst[[arg]] <- glbObsAll[, arg]; print(head(args_lst[[arg]])); print(head(drv_vals <- do.call(glbFeatsDerive[[tst]]$mapfn, args_lst))); 
# print(which_ix <- which(args_lst[[arg]] == 0.75)); print(drv_vals[which_ix]); 

glbFeatsDateTime <- list()
# Use OlsonNames() to enumerate supported time zones
# glbFeatsDateTime[["<DateTimeFeat>"]] <- 
#     c(format = "%Y-%m-%d %H:%M:%S" or "%m/%e/%y", timezone = "US/Eastern", impute.na = TRUE, 
#       last.ctg = FALSE, poly.ctg = FALSE)

glbFeatsPrice <- NULL # or c("<price_var>")

glbFeatsImage <- list() #list(<imageFeat> = list(patchSize = 10)) # if patchSize not specified, no patch computation

glbFeatsText <- list()
Sys.setlocale("LC_ALL", "C") # For english
## [1] "C/C/C/C/C/en_US.UTF-8"
#glbFeatsText[["<TextFeature>"]] <- list(NULL,
#   ,names = myreplacePunctuation(str_to_lower(gsub(" ", "", c(NULL, 
#       <comma-separated-screened-names>
#   ))))
#   ,rareWords = myreplacePunctuation(str_to_lower(gsub(" ", "", c(NULL, 
#       <comma-separated-nonSCOWL-words>
#   ))))
#)

# Text Processing Step: custom modifications not present in txt_munge -> use glbFeatsDerive
# Text Processing Step: universal modifications
glb_txt_munge_filenames_pfx <- "<projectId>_mytxt_"

# Text Processing Step: tolower
# Text Processing Step: myreplacePunctuation
# Text Processing Step: removeWords
glb_txt_stop_words <- list()
# Remember to use unstemmed words
if (length(glbFeatsText) > 0) {
    require(tm)
    require(stringr)

    glb_txt_stop_words[["<txt_var>"]] <- sort(myreplacePunctuation(str_to_lower(gsub(" ", "", c(NULL
        # Remove any words from stopwords            
#         , setdiff(myreplacePunctuation(stopwords("english")), c("<keep_wrd1>", <keep_wrd2>"))
                                
        # Remove salutations
        ,"mr","mrs","dr","Rev"                                

        # Remove misc
        #,"th" # Happy [[:digit::]]+th birthday 

        # Remove terms present in Trn only or New only; search for "Partition post-stem"
        #   ,<comma-separated-terms>        

        # cor.y.train == NA
#         ,unlist(strsplit(paste(c(NULL
#           ,"<comma-separated-terms>"
#         ), collapse=",")

        # freq == 1; keep c("<comma-separated-terms-to-keep>")
            # ,<comma-separated-terms>

        # chisq.pval high (e.g. == 1); keep c("<comma-separated-terms-to-keep>")
            # ,<comma-separated-terms>

        # nzv.freqRatio high (e.g. >= glbFeatsNzvFreqMax); keep c("<comma-separated-terms-to-keep>")
            # ,<comma-separated-terms>        
                                            )))))
}
#orderBy(~term, glb_post_stem_words_terms_df_lst[[txtFeat]][grep("^man", glb_post_stem_words_terms_df_lst[[txtFeat]]$term), ])
#glbObsAll[glb_post_stem_words_terms_mtrx_lst[[txtFeat]][, 4866] > 0, c(glb_rsp_var, txtFeat)]

# To identify terms with a specific freq
#paste0(sort(subset(glb_post_stop_words_terms_df_lst[[txtFeat]], freq == 1)$term), collapse = ",")
#paste0(sort(subset(glb_post_stem_words_terms_df_lst[[txtFeat]], freq <= 2)$term), collapse = ",")
#subset(glb_post_stem_words_terms_df_lst[[txtFeat]], term %in% c("zinger"))

# To identify terms with a specific freq & 
#   are not stemmed together later OR is value of color.fctr (e.g. gold)
#paste0(sort(subset(glb_post_stop_words_terms_df_lst[[txtFeat]], (freq == 1) & !(term %in% c("blacked","blemish","blocked","blocks","buying","cables","careful","carefully","changed","changing","chargers","cleanly","cleared","connect","connects","connected","contains","cosmetics","default","defaulting","defective","definitely","describe","described","devices","displays","drop","drops","engravement","excellant","excellently","feels","fix","flawlessly","frame","framing","gentle","gold","guarantee","guarantees","handled","handling","having","install","iphone","iphones","keeped","keeps","known","lights","line","lining","liquid","liquidation","looking","lots","manuals","manufacture","minis","most","mostly","network","networks","noted","opening","operated","performance","performs","person","personalized","photograph","physically","placed","places","powering","pre","previously","products","protection","purchasing","returned","rotate","rotation","running","sales","second","seconds","shipped","shuts","sides","skin","skinned","sticker","storing","thats","theres","touching","unusable","update","updates","upgrade","weeks","wrapped","verified","verify") ))$term), collapse = ",")

#print(subset(glb_post_stem_words_terms_df_lst[[txtFeat]], (freq <= 2)))
#glbObsAll[which(terms_mtrx[, 229] > 0), glbFeatsText]

# To identify terms with cor.y == NA
#orderBy(~-freq+term, subset(glb_post_stop_words_terms_df_lst[[txtFeat]], is.na(cor.y)))
#paste(sort(subset(glb_post_stop_words_terms_df_lst[[txtFeat]], is.na(cor.y))[, "term"]), collapse=",")
#orderBy(~-freq+term, subset(glb_post_stem_words_terms_df_lst[[txtFeat]], is.na(cor.y)))

# To identify terms with low cor.y.abs
#head(orderBy(~cor.y.abs+freq+term, subset(glb_post_stem_words_terms_df_lst[[txtFeat]], !is.na(cor.y))), 5)

# To identify terms with high chisq.pval
#subset(glb_post_stem_words_terms_df_lst[[txtFeat]], chisq.pval > 0.99)
#paste0(sort(subset(glb_post_stem_words_terms_df_lst[[txtFeat]], (chisq.pval > 0.99) & (freq <= 10))$term), collapse=",")
#paste0(sort(subset(glb_post_stem_words_terms_df_lst[[txtFeat]], (chisq.pval > 0.9))$term), collapse=",")
#head(orderBy(~-chisq.pval+freq+term, glb_post_stem_words_terms_df_lst[[txtFeat]]), 5)
#glbObsAll[glb_post_stem_words_terms_mtrx_lst[[txtFeat]][, 68] > 0, glbFeatsText]
#orderBy(~term, glb_post_stem_words_terms_df_lst[[txtFeat]][grep("^m", glb_post_stem_words_terms_df_lst[[txtFeat]]$term), ])

# To identify terms with high nzv.freqRatio
#summary(glb_post_stem_words_terms_df_lst[[txtFeat]]$nzv.freqRatio)
#paste0(sort(setdiff(subset(glb_post_stem_words_terms_df_lst[[txtFeat]], (nzv.freqRatio >= glbFeatsNzvFreqMax) & (freq < 10) & (chisq.pval >= 0.05))$term, c( "128gb","3g","4g","gold","ipad1","ipad3","ipad4","ipadair2","ipadmini2","manufactur","spacegray","sprint","tmobil","verizon","wifion"))), collapse=",")

# To identify obs with a txt term
#tail(orderBy(~-freq+term, glb_post_stop_words_terms_df_lst[[txtFeat]]), 20)
#mydspObs(list(descr.my.contains="non"), cols=c("color", "carrier", "cellular", "storage"))
#grep("ever", dimnames(terms_stop_mtrx)$Terms)
#which(terms_stop_mtrx[, grep("ipad", dimnames(terms_stop_mtrx)$Terms)] > 0)
#glbObsAll[which(terms_stop_mtrx[, grep("16", dimnames(terms_stop_mtrx)$Terms)[1]] > 0), c(glbFeatsCategory, "storage", txtFeat)]

# Text Processing Step: screen for names # Move to glbFeatsText specs section in order of text processing steps
# glbFeatsText[["<txtFeat>"]]$names <- myreplacePunctuation(str_to_lower(gsub(" ", "", c(NULL
#         # Person names for names screening
#         ,<comma-separated-list>
#         
#         # Company names
#         ,<comma-separated-list>
#                     
#         # Product names
#         ,<comma-separated-list>
#     ))))

# glbFeatsText[["<txtFeat>"]]$rareWords <- myreplacePunctuation(str_to_lower(gsub(" ", "", c(NULL
#         # Words not in SCOWL db
#         ,<comma-separated-list>
#     ))))

# To identify char vectors post glbFeatsTextMap
#grep("six(.*)hour", glb_txt_chr_lst[[txtFeat]], ignore.case = TRUE, value = TRUE)
#grep("[S|s]ix(.*)[H|h]our", glb_txt_chr_lst[[txtFeat]], value = TRUE)

# To identify whether terms shd be synonyms
#orderBy(~term, glb_post_stop_words_terms_df_lst[[txtFeat]][grep("^moder", glb_post_stop_words_terms_df_lst[[txtFeat]]$term), ])
# term_row_df <- glb_post_stop_words_terms_df_lst[[txtFeat]][grep("^came$", glb_post_stop_words_terms_df_lst[[txtFeat]]$term), ]
# 
# cor(glb_post_stop_words_terms_mtrx_lst[[txtFeat]][glbObsAll$.lcn == "Fit", term_row_df$pos], glbObsTrn[, glb_rsp_var], use="pairwise.complete.obs")

# To identify which stopped words are "close" to a txt term
#sort(glbFeatsCluster)

# Text Processing Step: stemDocument
# To identify stemmed txt terms
#glb_post_stop_words_terms_df_lst[[txtFeat]][grep("^la$", glb_post_stop_words_terms_df_lst[[txtFeat]]$term), ]
#orderBy(~term, glb_post_stem_words_terms_df_lst[[txtFeat]][grep("^con", glb_post_stem_words_terms_df_lst[[txtFeat]]$term), ])
#glbObsAll[which(terms_stem_mtrx[, grep("use", dimnames(terms_stem_mtrx)$Terms)[[1]]] > 0), c(glbFeatsId, "productline", txtFeat)]
#glbObsAll[which(TfIdf_stem_mtrx[, 191] > 0), c(glbFeatsId, glbFeatsCategory, txtFeat)]
#glbObsAll[which(glb_post_stop_words_terms_mtrx_lst[[txtFeat]][, 6165] > 0), c(glbFeatsId, glbFeatsCategory, txtFeat)]
#which(glbObsAll$UniqueID %in% c(11915, 11926, 12198))

# Text Processing Step: mycombineSynonyms
#   To identify which terms are associated with not -> combine "could not" & "couldn't"
#findAssocs(glb_full_DTM_lst[[txtFeat]], "not", 0.05)
#   To identify which synonyms should be combined
#orderBy(~term, glb_post_stem_words_terms_df_lst[[txtFeat]][grep("^c", glb_post_stem_words_terms_df_lst[[txtFeat]]$term), ])
chk_comb_cor <- function(syn_lst) {
#     cor(terms_stem_mtrx[glbObsAll$.src == "Train", grep("^(damag|dent|ding)$", dimnames(terms_stem_mtrx)[[2]])], glbObsTrn[, glb_rsp_var], use="pairwise.complete.obs")
    print(subset(glb_post_stem_words_terms_df_lst[[txtFeat]], term %in% syn_lst$syns))
    print(subset(get_corpus_terms(tm_map(glbFeatsTextCorpus[[txtFeat]], mycombineSynonyms, list(syn_lst), lazy=FALSE)), term == syn_lst$word))
#     cor(terms_stop_mtrx[glbObsAll$.src == "Train", grep("^(damage|dent|ding)$", dimnames(terms_stop_mtrx)[[2]])], glbObsTrn[, glb_rsp_var], use="pairwise.complete.obs")
#     cor(rowSums(terms_stop_mtrx[glbObsAll$.src == "Train", grep("^(damage|dent|ding)$", dimnames(terms_stop_mtrx)[[2]])]), glbObsTrn[, glb_rsp_var], use="pairwise.complete.obs")
}
#chk_comb_cor(syn_lst=list(word="cabl",  syns=c("cabl", "cord")))
#chk_comb_cor(syn_lst=list(word="damag",  syns=c("damag", "dent", "ding")))
#chk_comb_cor(syn_lst=list(word="dent",  syns=c("dent", "ding")))
#chk_comb_cor(syn_lst=list(word="use",  syns=c("use", "usag")))

glbFeatsTextSynonyms <- list()
# list parsed to collect glbFeatsText[[<txtFeat>]]$vldTerms
# glbFeatsTextSynonyms[["Hdln.my"]] <- list(NULL
#     # people in places
#     , list(word = "australia", syns = c("australia", "australian"))
#     , list(word = "italy", syns = c("italy", "Italian"))
#     , list(word = "newyork", syns = c("newyork", "newyorker"))    
#     , list(word = "Pakistan", syns = c("Pakistan", "Pakistani"))    
#     , list(word = "peru", syns = c("peru", "peruvian"))
#     , list(word = "qatar", syns = c("qatar", "qatari"))
#     , list(word = "scotland", syns = c("scotland", "scotish"))
#     , list(word = "Shanghai", syns = c("Shanghai", "Shanzhai"))    
#     , list(word = "venezuela", syns = c("venezuela", "venezuelan"))    
# 
#     # companies - needs to be data dependent 
#     #   - e.g. ensure BNP in this experiment/feat always refers to BNPParibas
#         
#     # general synonyms
#     , list(word = "Create", syns = c("Create","Creator")) 
#     , list(word = "cute", syns = c("cute","cutest"))     
#     , list(word = "Disappear", syns = c("Disappear","Fadeout"))     
#     , list(word = "teach", syns = c("teach", "taught"))     
#     , list(word = "theater",  syns = c("theater", "theatre", "theatres")) 
#     , list(word = "understand",  syns = c("understand", "understood"))    
#     , list(word = "weak",  syns = c("weak", "weaken", "weaker", "weakest"))
#     , list(word = "wealth",  syns = c("wealth", "wealthi"))    
#     
#     # custom synonyms (phrases)
#     
#     # custom synonyms (names)
#                                       )
#glbFeatsTextSynonyms[["<txtFeat>"]] <- list(NULL
#     , list(word="<stem1>",  syns=c("<stem1>", "<stem1_2>"))
#                                       )

for (txtFeat in names(glbFeatsTextSynonyms))
    for (entryIx in 1:length(glbFeatsTextSynonyms[[txtFeat]])) {
        glbFeatsTextSynonyms[[txtFeat]][[entryIx]]$word <-
            str_to_lower(glbFeatsTextSynonyms[[txtFeat]][[entryIx]]$word)
        glbFeatsTextSynonyms[[txtFeat]][[entryIx]]$syns <-
            str_to_lower(glbFeatsTextSynonyms[[txtFeat]][[entryIx]]$syns)        
    }        

glbFeatsTextSeed <- 181
# tm options include: check tm::weightSMART 
glb_txt_terms_control <- list( # Gather model performance & run-time stats
                    # weighting = function(x) weightSMART(x, spec = "nnn")
                    # weighting = function(x) weightSMART(x, spec = "lnn")
                    # weighting = function(x) weightSMART(x, spec = "ann")
                    # weighting = function(x) weightSMART(x, spec = "bnn")
                    # weighting = function(x) weightSMART(x, spec = "Lnn")
                    # 
                    weighting = function(x) weightSMART(x, spec = "ltn") # default
                    # weighting = function(x) weightSMART(x, spec = "lpn")                    
                    # 
                    # weighting = function(x) weightSMART(x, spec = "ltc")                    
                    # 
                    # weighting = weightBin 
                    # weighting = weightTf 
                    # weighting = weightTfIdf # : default
                # termFreq selection criteria across obs: tm default: list(global=c(1, Inf))
                    , bounds = list(global = c(1, Inf)) 
                # wordLengths selection criteria: tm default: c(3, Inf)
                    , wordLengths = c(1, Inf) 
                              ) 

glb_txt_cor_var <- glb_rsp_var # : default # or c(<feat>)

# select one from c("union.top.val.cor", "top.cor", "top.val", default: "top.chisq", "sparse")
glbFeatsTextFilter <- "top.chisq" 
glbFeatsTextTermsMax <- rep(10, length(glbFeatsText)) # :default
names(glbFeatsTextTermsMax) <- names(glbFeatsText)

# Text Processing Step: extractAssoc
glbFeatsTextAssocCor <- rep(1, length(glbFeatsText)) # :default 
names(glbFeatsTextAssocCor) <- names(glbFeatsText)

# Remember to use stemmed terms
glb_important_terms <- list()

# Text Processing Step: extractPatterns (ngrams)
glbFeatsTextPatterns <- list()
#glbFeatsTextPatterns[[<txtFeat>>]] <- list()
#glbFeatsTextPatterns[[<txtFeat>>]] <- c(metropolitan.diary.colon = "Metropolitan Diary:")

# Have to set it even if it is not used
# Properties:
#   numrows(glb_feats_df) << numrows(glbObsFit
#   Select terms that appear in at least 0.2 * O(FP/FN(glbObsOOB)) ???
#       numrows(glbObsOOB) = 1.1 * numrows(glbObsNew) ???
glb_sprs_thresholds <- NULL # or c(<txtFeat1> = 0.988, <txtFeat2> = 0.970, <txtFeat3> = 0.970)

glbFctrMaxUniqVals <- 20 # default: 20
glb_impute_na_data <- FALSE # or TRUE
glb_mice_complete.seed <- 144 # or any integer

glbFeatsCluster <- paste(grep("^Q.", glbFeatsExclude, value = TRUE), "fctr", sep = ".") # NULL : glbFeatsCluster <- c("YOB.Age.fctr", "Gender.fctr", "Income.fctr", 
                     # # "Hhold.fctr",
                     # "Edn.fctr",
                     # paste(grep("^Q.", glbFeatsExclude, value = TRUE), "fctr", sep = ".")) # NULL : default or c("<feat1>", "<feat2>")
# glbFeatsCluster <- grep(paste0("[", 
#                         toupper(paste0(substr(glbFeatsText, 1, 1), collapse = "")),
#                                       "]\\.[PT]\\."), 
#                                names(glbObsAll), value = TRUE)

glb_cluster.seed <- 189 # or any integer
glbClusterEntropyVar <- NULL # c(glb_rsp_var, as.factor(cut(glb_rsp_var, 3)), default: NULL)
glbFeatsClusterVarsExclude <- FALSE # default FALSE

glb_interaction_only_feats <- NULL # : default or c(<parent_feat> = "<child_feat>")

glbFeatsNzvFreqMax <- 19 # 19 : caret default
glbFeatsNzvUniqMin <- 10 # 10 : caret default

glbRFESizes <- list()
#glbRFESizes[["mdlFamily"]] <- c(4, 8, 16, 32, 64, 67, 68, 69) # Accuracy@69/70 = 0.8258
# glbRFESizes[["RFE.X"]] <- c(4, 6, 7, 8, 9, 10, 16, 32, 64, 128, 253) # accuracy(8) = 0.5648
# glbRFESizes[["Final"]] <- c(8, 16, 32, 40, 44, 46, 48, 49, 50, 51, 52, 56, 64, 96, 128, 247) # accuracy(49) = 0.6164

glbRFEResults <- NULL

glbObsFitOutliers <- list()
# If outliers.n >= 10; consider concatenation of interaction vars
# glbObsFitOutliers[["<mdlFamily>"]] <- c(NULL
#     is.na(.rstudent)
#     max(.rstudent)
#     is.na(.dffits)
#     .hatvalues >= 0.99        
#     -38,167,642 < minmax(.rstudent) < 49,649,823    
#     , <comma-separated-<glbFeatsId>>
#                                     )
glbObsTrnOutliers <- list()
glbObsTrnOutliers[["Final"]] <- union(glbObsFitOutliers[["All.X"]],
                                c(NULL
                                ))

# Modify mdlId to (build & extract) "<FamilyId>#<Fit|Trn>#<caretMethod>#<preProc1.preProc2>#<samplingMethod>"
glb_models_lst <- list(); glb_models_df <- data.frame()

# Add xgboost algorithm

# Regression
if (glb_is_regression) {
    glbMdlMethods <- c(NULL
        # deterministic
            #, "lm", # same as glm
            , "glm", "bayesglm", "glmnet"
            , "rpart"
        # non-deterministic
            , "gbm", "rf" 
        # Unknown
            , "nnet" , "avNNet" # runs 25 models per cv sample for tunelength=5
            , "svmLinear", "svmLinear2"
            , "svmPoly" # runs 75 models per cv sample for tunelength=5
            , "svmRadial" 
            , "earth"
            , "bagEarth" # Takes a long time
            ,"xgbLinear","xgbTree"
        )
} else
# Classification - Add ada (auto feature selection)
    if (glb_is_binomial)
        glbMdlMethods <- c(NULL
        # deterministic                     
            , "bagEarth" # Takes a long time        
            , "glm", "bayesglm", "glmnet"
            , "nnet"
            , "rpart"
        # non-deterministic        
            , "gbm"
            , "avNNet" # runs 25 models per cv sample for tunelength=5      
            , "rf"
        # Unknown
            , "lda", "lda2"
                # svm models crash when predict is called -> internal to kernlab it should call predict without .outcome
            , "svmLinear", "svmLinear2"
            , "svmPoly" # runs 75 models per cv sample for tunelength=5
            , "svmRadial" 
            , "earth"
            ,"xgbLinear","xgbTree"
        ) else
        glbMdlMethods <- c(NULL
        # deterministic
            ,"glmnet"
        # non-deterministic 
            ,"rf"       
        # Unknown
            ,"gbm","rpart","xgbLinear","xgbTree"
        )

glbMdlFamilies <- list(); glb_mdl_feats_lst <- list()
# family: Choose from c("RFE.X", "Csm.X", "All.X", "Best.Interact") %*% c(NUll, ".NOr", ".Inc")
#   RFE = "Recursive Feature Elimination"
#   Csm = CuStoM
#   NOr = No OutlieRs
#   Inc = INteraCt
#   methods: Choose from c(NULL, <method>, glbMdlMethods) 
#glbMdlFamilies[["RFE.X"]] <- c("glmnet", "glm") # non-NULL vector is mandatory
if (glb_is_classification && !glb_is_binomial) {
    # glm does not work for multinomial
    glbMdlFamilies[["All.X"]] <- c("glmnet") 
} else {
    # glbMdlFamilies[["All.X"]] <- c("glmnet", "glm")
    glbMdlFamilies[["All.X"]] <- c("glmnet")
    # glbMdlFamilies[["All.X"]] <- setdiff(glbMdlMethods, c(NULL
    #     # , "bayesglm" # error: Error in trControl$classProbs && any(classLevels != make.names(classLevels)) : invalid 'x' type in 'x && y'
    #     , "lda" # error: model fit failed for Fold1.Rep1: parameter=none Error in lda.default(x, grouping, ...)
    #     ,"lda2" # error: There were missing values in resampled performance measures.
    #     , "svmLinear" # Error in .local(object, ...) : test vector does not match model ! In addition: Warning messages:
    #     , "svmLinear2" # SVM has not been trained using `probability = TRUE`, probabilities not available for predictions
    #     , "svmPoly" # runs 75 models per cv sample for tunelength=5 # took > 2 hrs # Error in .local(object, ...) : test vector does not match model !
    #     , "svmRadial" # Error in .local(object, ...) : test vector does not match model !
    #     ,"xgbLinear","xgbTree" # Need clang-omp compiler; Upgrade to Revolution R 3.2.3 (3.2.2 current); https://github.com/dmlc/xgboost/issues/276 thread
    #                                     ))

    # glbMdlFamilies[["RFE.X"]] <- c("glmnet", "glm")
    # glbMdlFamilies[["RFE.X"]] <- c("glmnet")
    # glbMdlFamilies[["RFE.X"]] <- setdiff(glbMdlMethods, c(NULL
    #     # , "bayesglm" # error: Error in trControl$classProbs && any(classLevels != make.names(classLevels)) : invalid 'x' type in 'x && y'
    #     # , "lda","lda2" # error: Error in lda.default(x, grouping, ...) : variable 236 appears to be constant within groups
    #     , "svmLinear" # Error in .local(object, ...) : test vector does not match model ! In addition: Warning messages:
    #     , "svmLinear2" # SVM has not been trained using `probability = TRUE`, probabilities not available for predictions
    #     , "svmPoly" # runs 75 models per cv sample for tunelength=5 # took > 2 hrs # Error in .local(object, ...) : test vector does not match model !
    #     , "svmRadial" # Error in .local(object, ...) : test vector does not match model !
    #     ,"xgbLinear","xgbTree" # Need clang-omp compiler; Upgrade to Revolution R 3.2.3 (3.2.2 current); https://github.com/dmlc/xgboost/issues/276 thread
    #                                     ))
}
# glbMdlFamilies[["All.X.Inc"]] <- glbMdlFamilies[["All.X"]] # value not used
# glbMdlFamilies[["RFE.X.Inc"]] <- glbMdlFamilies[["RFE.X"]] # value not used

# Check if interaction features make RFE better
# glbMdlFamilies[["CSM.X"]] <- setdiff(glbMdlMethods, c("lda", "lda2")) # crashing due to category:.clusterid ??? #c("glmnet", "glm") # non-NULL list is mandatory
# glb_mdl_feats_lst[["CSM.X"]] <- c(NULL
#     , <comma-separated-features-vector>
#                                   )
# dAFeats.CSM.X %<d-% c(NULL
#     # Interaction feats up to varImp(RFE.X.glmnet) >= 50
#     , <comma-separated-features-vector>
#     , setdiff(myextract_actual_feats(predictors(glbRFEResults)), c(NULL
#                , <comma-separated-features-vector>
#                                                                       ))    
#                                   )
# glb_mdl_feats_lst[["CSM.X"]] <- "%<d-% dAFeats.CSM.X"

# glbMdlFamilies[["Final"]] <- c(NULL) # NULL vector acceptable # c("glmnet", "glm")

glbMdlAllowParallel <- list()
#glbMdlAllowParallel[["Final##rcv#glmnet"]] <- FALSE

# Check if tuning parameters make fit better; make it mdlFamily customizable ?
glbMdlTuneParams <- data.frame()

# When glmnet crashes at model$grid with error: ???
# AllX__rcv_glmnetTuneParams <- rbind(data.frame() # alpha shd be <= 1.0 ALWAYS
#     ,data.frame(parameter = "alpha",  vals = "0.100 0.325 0.550 0.775 1.000")
#     ,data.frame(parameter = "lambda", vals = "0.0053781495 0.01 0.0249631588 0.03 0.04454817")
#                         ) # max.Accuracy.OOB = 0.5981941 @ 0.775 0.02496316
# AllX_YeoJohnson_rcv_glmnetTuneParams <- rbind(data.frame()
#     ,data.frame(parameter = "alpha",  vals = "0.100 0.325 0.550 0.775 1.000")
#     ,data.frame(parameter = "lambda", vals = "0.0053781495 0.01 0.0249631588 0.03 0.04454817")
#                         ) # max.Accuracy.OOB = 0.6004515 @ 0.775 0.02496316
# AllX_zvpca_rcv_glmnetTuneParams <- rbind(data.frame()
#     ,data.frame(parameter = "alpha",  vals = "0.325 0.550 0.775 0.9 1.000")
#     ,data.frame(parameter = "lambda", vals = "1.040899e-03 0.003 4.831424e-03 0.01 2.242548e-02")
#                         ) # max.Accuracy.OOB = 0.6185102 @ 1.0 0.004831424
#                           #                    0.616408  @ 0.9 0.004831424
# 
# glbMdlTuneParams <- rbind(glbMdlTuneParams
#     ,cbind(data.frame(mdlId = "All.X##rcv#glmnet"),         AllX__rcv_glmnetTuneParams)
#     ,cbind(data.frame(mdlId = "All.X#YeoJohnson#rcv#glmnet"), 
#                                                             AllX_YeoJohnson_rcv_glmnetTuneParams)      ,cbind(data.frame(mdlId = "All.X#zv.pca#rcv#glmnet"), 
#                                                             AllX_zvpca_rcv_glmnetTuneParams)   
# )

    #avNNet    
    #   size=[1] 3 5 7 9; decay=[0] 1e-04 0.001  0.01   0.1; bag=[FALSE]; RMSE=1.3300906 

    #bagEarth
    #   degree=1 [2] 3; nprune=64 128 256 512 [1024]; RMSE=0.6486663 (up)
# bagEarthTuneParams <- rbind(data.frame()
#                         ,data.frame(parameter = "degree", vals = "1")
#                         ,data.frame(parameter = "nprune", vals = "256")
#                         )
# glbMdlTuneParams <- myrbind_df(glbMdlTuneParams,
#                                cbind(data.frame(mdlId = "Final.RFE.X.Inc##rcv#bagEarth"),
#                                      bagEarthTuneParams))

# glbMdlTuneParams <- myrbind_df(glbMdlTuneParams, rbind(data.frame()
#     ,data.frame(method = "bagEarth", parameter = "nprune", vals = "256")
#     ,data.frame(method = "bagEarth", parameter = "degree", vals = "2")    
# ))

    #earth 
    #   degree=[1]; nprune=2  [9] 17 25 33; RMSE=0.1334478
    
    #gbm 
    #   shrinkage=0.05 [0.10] 0.15 0.20 0.25; n.trees=100 150 200 [250] 300; interaction.depth=[1] 2 3 4 5; n.minobsinnode=[10]; RMSE=0.2008313     
# glbMdlTuneParams <- myrbind_df(glbMdlTuneParams, rbind(data.frame()
#     ,data.frame(method = "gbm", parameter = "shrinkage", min = 0.05, max = 0.25, by = 0.05)
#     ,data.frame(method = "gbm", parameter = "n.trees", min = 100, max = 300, by = 50)
#     ,data.frame(method = "gbm", parameter = "interaction.depth", min = 1, max = 5, by = 1)
#     ,data.frame(method = "gbm", parameter = "n.minobsinnode", min = 10, max = 10, by = 10)
#     #seq(from=0.05,  to=0.25, by=0.05)
# ))

    #glmnet
    #   alpha=0.100 [0.325] 0.550 0.775 1.000; lambda=0.0005232693 0.0024288010 0.0112734954 [0.0523269304] 0.2428800957; RMSE=0.6164891
# glbMdlTuneParams <- myrbind_df(glbMdlTuneParams, rbind(data.frame()
#     ,data.frame(method = "glmnet", parameter = "alpha", vals = "0.550 0.775 0.8875 0.94375 1.000")
#     ,data.frame(method = "glmnet", parameter = "lambda", vals = "9.858855e-05 0.0001971771 0.0009152152 0.0042480525 0.0197177130")    
# ))

    #nnet    
    #   size=3 5 [7] 9 11; decay=0.0001 0.001 0.01 [0.1] 0.2; RMSE=0.9287422
# glbMdlTuneParams <- myrbind_df(glbMdlTuneParams, rbind(data.frame()
#     ,data.frame(method = "nnet", parameter = "size", vals = "3 5 7 9 11")
#     ,data.frame(method = "nnet", parameter = "decay", vals = "0.0001 0.0010 0.0100 0.1000 0.2000")    
# ))

    #rf # Don't bother; results are not deterministic
    #       mtry=2  35  68 [101] 134; RMSE=0.1339974
# glbMdlTuneParams <- myrbind_df(glbMdlTuneParams, rbind(data.frame()
#     ,data.frame(method = "rf", parameter = "mtry", vals = "2 5 9 13 17")
# ))

    #rpart 
    #   cp=0.020 [0.025] 0.030 0.035 0.040; RMSE=0.1770237
# glbMdlTuneParams <- myrbind_df(glbMdlTuneParams, rbind(data.frame()    
#     ,data.frame(method = "rpart", parameter = "cp", vals = "0.004347826 0.008695652 0.017391304 0.021739130 0.034782609")
# ))
    
    #svmLinear
    #   C=0.01 0.05 [0.10] 0.50 1.00 2.00 3.00 4.00; RMSE=0.1271318; 0.1296718
# glbMdlTuneParams <- myrbind_df(glbMdlTuneParams, rbind(data.frame()
#     ,data.frame(method = "svmLinear", parameter = "C", vals = "0.01 0.05 0.1 0.5 1")
# ))

    #svmLinear2    
    #   cost=0.0625 0.1250 [0.25] 0.50 1.00; RMSE=0.1276354 
# glbMdlTuneParams <- myrbind_df(glbMdlTuneParams, rbind(data.frame()
#     ,data.frame(method = "svmLinear2", parameter = "cost", vals = "0.0625 0.125 0.25 0.5 1")
# ))

    #svmPoly    
    #   degree=[1] 2 3 4 5; scale=0.01 0.05 [0.1] 0.5 1; C=0.50 1.00 [2.00] 3.00 4.00; RMSE=0.1276130
# glbMdlTuneParams <- myrbind_df(glbMdlTuneParams, rbind(data.frame()
#     ,data.frame(method="svmPoly", parameter="degree", min=1, max=5, by=1) #seq(1, 5, 1)
#     ,data.frame(method="svmPoly", parameter="scale", vals="0.01, 0.05, 0.1, 0.5, 1")
#     ,data.frame(method="svmPoly", parameter="C", vals="0.50, 1.00, 2.00, 3.00, 4.00")    
# ))

    #svmRadial
    #   sigma=[0.08674323]; C=0.25 0.50 1.00 [2.00] 4.00; RMSE=0.1614957
    
#glb2Sav(); all.equal(sav_models_df, glb_models_df)

pkgPreprocMethods <-     
# caret version: 6.0.068 # packageVersion("caret")
# operations are applied in this order: zero-variance filter, near-zero variance filter, Box-Cox/Yeo-Johnson/exponential transformation, centering, scaling, range, imputation, PCA, ICA then spatial sign
# *Impute methods needed only if NAs are fed to myfit_mdl
#   Also, ordered.factor in caret creates features as Edn.fctr^4 which is treated as an exponent by bagImpute
    c(NULL
      ,"zv", "nzv"
      ,"BoxCox", "YeoJohnson", "expoTrans"
      ,"center", "scale", "center.scale", "range"
      ,"knnImpute", "bagImpute", "medianImpute"
      ,"zv.pca", "ica", "spatialSign"
      ,"conditionalX") 

glbMdlPreprocMethods <- list(NULL # NULL # : default
    # ,"All.X" = list("glmnet" = union(setdiff(pkgPreprocMethods,
    #                                         c("knnImpute", "bagImpute", "medianImpute")),
    #                                 # c(NULL)))
    #                                 c("zv.YeoJohnson.pca")))
    # ,"RFE.X" = list("glmnet" = union(setdiff(pkgPreprocMethods,
    #                                         c("knnImpute", "bagImpute", "medianImpute")),
    #                                 c(NULL)))
    #                                 # c("zv.pca.spatialSign")))
)
# glbMdlPreprocMethods[["RFE.X"]] <- list("glmnet" = union(unlist(glbMdlPreprocMethods[["All.X"]]),
#                                                     "nzv.pca.spatialSign"))

# Baseline prediction model feature(s)
glb_Baseline_mdl_var <- NULL # or c("<feat>")

glbMdlMetric_terms <- NULL # or matrix(c(
#                               0,1,2,3,4,
#                               2,0,1,2,3,
#                               4,2,0,1,2,
#                               6,4,2,0,1,
#                               8,6,4,2,0
#                           ), byrow=TRUE, nrow=5)
glbMdlMetricSummary <- NULL # or "<metric_name>"
glbMdlMetricMaximize <- NULL # or FALSE (TRUE is not the default for both classification & regression) 
glbMdlMetricSummaryFn <- NULL # or function(data, lev=NULL, model=NULL) {
#     confusion_mtrx <- t(as.matrix(confusionMatrix(data$pred, data$obs)))
#     #print(confusion_mtrx)
#     #print(confusion_mtrx * glbMdlMetric_terms)
#     metric <- sum(confusion_mtrx * glbMdlMetric_terms) / nrow(data)
#     names(metric) <- glbMdlMetricSummary
#     return(metric)
# }

glbMdlCheckRcv <- FALSE # Turn it on when needed; otherwise takes long time
glb_rcv_n_folds <- 3 # or NULL
glb_rcv_n_repeats <- 3 # or NULL

glb_clf_proba_threshold <- NULL # 0.5

# Model selection criteria
if (glb_is_regression)
    glbMdlMetricsEval <- c("min.RMSE.OOB", "max.R.sq.OOB", "min.elapsedtime.everything",
                           "max.Adj.R.sq.fit", "min.RMSE.fit")
    #glbMdlMetricsEval <- c("min.RMSE.fit", "max.R.sq.fit", "max.Adj.R.sq.fit")    
if (glb_is_classification) {
    if (glb_is_binomial)
        glbMdlMetricsEval <- 
            c("max.Accuracy.OOB", "max.AUCROCR.OOB", "max.AUCpROC.OOB",
              "min.elapsedtime.everything", 
              # "min.aic.fit", 
              "max.Accuracy.fit") else        
        glbMdlMetricsEval <- c("max.Accuracy.OOB", "max.Kappa.OOB", "min.elapsedtime.everything")
}

# select from NULL [no ensemble models], "auto" [all models better than MFO or Baseline], c(mdl_ids in glb_models_lst) [Typically top-rated models in auto]
glbMdlEnsemble <- "auto" # NULL : default #"auto"
#     "%<d-% setdiff(mygetEnsembleAutoMdlIds(), 'CSM.X.rf')" 
#     c(<comma-separated-mdlIds>
#      )
glbMdlEnsembleSampleMethods <- c("boot", "boot632", "cv", "repeatedcv"
               # , "LOOCV" # tuneLength * nrow(fitDF) # way too many models
               , "LGOCV"
               , "adaptive_cv" # crashed for Q109244No
               # , "adaptive_boot"  #error: adaptive$min should be less than 3
               # , "adaptive_LGOCV" #error: adaptive$min should be less than 3
               )


# Only for classifications; for regressions remove "(.*)\\.prob" form the regex
# tmp_fitobs_df <- glbObsFit[, grep(paste0("^", gsub(".", "\\.", mygetPredictIds$value, fixed = TRUE), "CSM\\.X\\.(.*)\\.prob"), names(glbObsFit), value = TRUE)]; cor_mtrx <- cor(tmp_fitobs_df); cor_vctr <- sort(cor_mtrx[row.names(orderBy(~-Overall, varImp(glb_models_lst[["Ensemble.repeatedcv.glmnet"]])$imp))[1], ]); summary(cor_vctr); cor_vctr
#ntv.glm <- glm(reformulate(indepVar, glb_rsp_var), family = "binomial", data = glbObsFit)
#step.glm <- step(ntv.glm)

glbMdlSelId <- NULL #select from c(NULL, "All.X##rcv#glmnet", "RFE.X##rcv#glmnet", <mdlId>)
glbMdlFinId <- NULL #select from c(NULL, glbMdlSelId)

glb_dsp_cols <- c(".pos", glbFeatsId, glbFeatsCategory, glb_rsp_var
#               List critical cols excl. above
                  )

# Output specs
# lclgetfltout_df <- function(obsOutFinDf) {
#     require(tidyr)
#     obsOutFinDf <- obsOutFinDf %>%
#         tidyr::separate("ImageId.x.y", c(".src", ".pos", "x", "y"), 
#                         sep = "#", remove = TRUE, extra = "merge")
#     # mnm prefix stands for max_n_mean
#     mnmout_df <- obsOutFinDf %>%
#         dplyr::group_by(.pos) %>%
#         #dplyr::top_n(1, Probability1) %>% # Score = 3.9426         
#         #dplyr::top_n(2, Probability1) %>% # Score = ???; weighted = 3.94254;         
#         #dplyr::top_n(3, Probability1) %>% # Score = 3.9418; weighted = 3.94169; 
#         dplyr::top_n(4, Probability1) %>% # Score = ???; weighted = 3.94149;        
#         #dplyr::top_n(5, Probability1) %>% # Score = 3.9421; weighted = 3.94178
#     
#         # dplyr::summarize(xMeanN = mean(as.numeric(x)), yMeanN = mean(as.numeric(y)))
#         # dplyr::summarize(xMeanN = weighted.mean(as.numeric(x), Probability1), yMeanN = mean(as.numeric(y)))
#         # dplyr::summarize(xMeanN = weighted.mean(as.numeric(x), c(Probability1, 0.2357323, 0.2336925)), yMeanN = mean(as.numeric(y)))    
#         # dplyr::summarize(xMeanN = weighted.mean(as.numeric(x), c(Probability1)), yMeanN = mean(as.numeric(y)))
#         dplyr::summarize(xMeanN = weighted.mean(as.numeric(x), c(Probability1)), 
#                          yMeanN = weighted.mean(as.numeric(y), c(Probability1)))  
#     
#     maxout_df <- obsOutFinDf %>%
#         dplyr::group_by(.pos) %>%
#         dplyr::summarize(maxProb1 = max(Probability1))
#     fltout_df <- merge(maxout_df, obsOutFinDf, 
#                        by.x = c(".pos", "maxProb1"), by.y = c(".pos", "Probability1"),
#                        all.x = TRUE)
#     fmnout_df <- merge(fltout_df, mnmout_df, 
#                        by.x = c(".pos"), by.y = c(".pos"),
#                        all.x = TRUE)
#     return(fmnout_df)
# }
glbObsOut <- list(NULL
        # glbFeatsId will be the first output column, by default
        ,vars = list()
#         ,mapFn = function(obsOutFinDf) {
#                   }
                  )
#obsOutFinDf <- savobsOutFinDf
# glbObsOut$mapFn <- function(obsOutFinDf) {
#     txfout_df <- dplyr::select(obsOutFinDf, -.pos.y) %>%
#         dplyr::mutate(
#             lunch     = levels(glbObsTrn[, "lunch"    ])[
#                        round(mean(as.numeric(glbObsTrn[, "lunch"    ])), 0)],
#             dinner    = levels(glbObsTrn[, "dinner"   ])[
#                        round(mean(as.numeric(glbObsTrn[, "dinner"   ])), 0)],
#             reserve   = levels(glbObsTrn[, "reserve"  ])[
#                        round(mean(as.numeric(glbObsTrn[, "reserve"  ])), 0)],
#             outdoor   = levels(glbObsTrn[, "outdoor"  ])[
#                        round(mean(as.numeric(glbObsTrn[, "outdoor"  ])), 0)],
#             expensive = levels(glbObsTrn[, "expensive"])[
#                        round(mean(as.numeric(glbObsTrn[, "expensive"])), 0)],
#             liquor    = levels(glbObsTrn[, "liquor"   ])[
#                        round(mean(as.numeric(glbObsTrn[, "liquor"   ])), 0)],
#             table     = levels(glbObsTrn[, "table"    ])[
#                        round(mean(as.numeric(glbObsTrn[, "table"    ])), 0)],
#             classy    = levels(glbObsTrn[, "classy"   ])[
#                        round(mean(as.numeric(glbObsTrn[, "classy"   ])), 0)],
#             kids      = levels(glbObsTrn[, "kids"     ])[
#                        round(mean(as.numeric(glbObsTrn[, "kids"     ])), 0)]
#                       )
#     
#     print("ObsNew output class tables:")
#     print(sapply(c("lunch","dinner","reserve","outdoor",
#                    "expensive","liquor","table",
#                    "classy","kids"), 
#                  function(feat) table(txfout_df[, feat], useNA = "ifany")))
#     
#     txfout_df <- txfout_df %>%
#         dplyr::mutate(labels = "") %>%
#         dplyr::mutate(labels = 
#     ifelse(lunch     != "-1", paste(labels, lunch    ), labels)) %>%
#         dplyr::mutate(labels = 
#     ifelse(dinner    != "-1", paste(labels, dinner   ), labels)) %>%
#         dplyr::mutate(labels = 
#     ifelse(reserve   != "-1", paste(labels, reserve  ), labels)) %>%
#         dplyr::mutate(labels = 
#     ifelse(outdoor   != "-1", paste(labels, outdoor  ), labels)) %>%
#         dplyr::mutate(labels =         
#     ifelse(expensive != "-1", paste(labels, expensive), labels)) %>%
#         dplyr::mutate(labels =         
#     ifelse(liquor    != "-1", paste(labels, liquor   ), labels)) %>%
#         dplyr::mutate(labels =         
#     ifelse(table     != "-1", paste(labels, table    ), labels)) %>%
#         dplyr::mutate(labels =         
#     ifelse(classy    != "-1", paste(labels, classy   ), labels)) %>%
#         dplyr::mutate(labels =         
#     ifelse(kids      != "-1", paste(labels, kids     ), labels)) %>%
#         dplyr::select(business_id, labels)
#     return(txfout_df)
# }
#if (!is.null(glbObsOut$mapFn)) obsOutFinDf <- glbObsOut$mapFn(obsOutFinDf); print(head(obsOutFinDf))

glb_out_obs <- NULL # select from c(NULL : default to "new", "all", "new", "trn")

if (glb_is_classification && glb_is_binomial) {
    # glbObsOut$vars[["Probability1"]] <- 
    #     "%<d-% glbObsNew[, mygetPredictIds(glb_rsp_var, glbMdlFinId)$prob]" 
    # glbObsOut$vars[[glb_rsp_var_raw]] <-
    #     "%<d-% glb_map_rsp_var_to_raw(glbObsNew[,
    #                                         mygetPredictIds(glb_rsp_var, glbMdlFinId)$value])"
    glbObsOut$vars[["Predictions"]] <-
        "%<d-% glb_map_rsp_var_to_raw(glbObsNew[,
                                            mygetPredictIds(glb_rsp_var, glbMdlFinId)$value])"
} else {
#     glbObsOut$vars[[glbFeatsId]] <- 
#         "%<d-% as.integer(gsub('Test#', '', glbObsNew[, glbFeatsId]))"
    glbObsOut$vars[[glb_rsp_var]] <- 
        "%<d-% glbObsNew[, mygetPredictIds(glb_rsp_var, glbMdlFinId)$value]"
#     for (outVar in setdiff(glbFeatsExcludeLcl, glb_rsp_var_raw))
#         glbObsOut$vars[[outVar]] <- 
#             paste0("%<d-% mean(glbObsAll[, \"", outVar, "\"], na.rm = TRUE)")
}    
# glbObsOut$vars[[glb_rsp_var_raw]] <- glb_rsp_var_raw
# glbObsOut$vars[[paste0(head(unlist(strsplit(mygetPredictIds$value, "")), -1), collapse = "")]] <-

glbOutStackFnames <- # NULL #: default
    c("Q109244NA_Ensemble_cnk03_rest_out_fin.csv") 
    # c("Q109244No_AllXpreProc_cnk03_rest_out_fin.csv") 
    # c("Votes_Ensemble_cnk06_out_fin.csv") 


glbOut <- list(pfx = "Q109244Yes_AllX_manage.missing.data_")
# lclImageSampleSeed <- 129
glbOutDataVizFname <- NULL # choose from c(NULL, "<projectId>_obsall.csv")


glbChunks <- list(labels = c("set_global_options_wd","set_global_options"
    ,"import.data","inspect.data","scrub.data","transform.data"
    ,"extract.features"
        ,"extract.features.datetime","extract.features.image","extract.features.price"
        ,"extract.features.text","extract.features.string"  
        ,"extract.features.end"
    ,"manage.missing.data","cluster.data","partition.data.training","select.features"
    ,"fit.models_0","fit.models_1","fit.models_2","fit.models_3"
    ,"fit.data.training_0","fit.data.training_1"
    ,"predict.data.new"         
    ,"display.session.info"))
# To ensure that all chunks in this script are in glbChunks
if (!is.null(chkChunksLabels <- knitr::all_labels()) && # knitr::all_labels() doesn't work in console runs
    !identical(chkChunksLabels, glbChunks$labels)) {
    print(sprintf("setdiff(chkChunksLabels, glbChunks$labels): %s", 
                  setdiff(chkChunksLabels, glbChunks$labels)))    
    print(sprintf("setdiff(glbChunks$labels, chkChunksLabels): %s", 
                  setdiff(glbChunks$labels, chkChunksLabels)))    
}

glbChunks[["first"]] <- NULL # NULL # default: script will load envir from previous chunk
glbChunks[["last" ]] <- "manage.missing.data" # default: script will save envir at end of this chunk 
glbChunks[["inpFilePathName"]] <- NULL #"data/Q109244NA_Ensemble_Prep_fit.models_1_fit.models_1.RData" # NULL: default or "data/<prvScriptName>_<lstChunkLbl>.RData"
#mysavChunk(glbOut$pfx, glbChunks[["last"]]) # called from myevlChunk
# Temporary: Delete this function (if any) from here after appropriate .RData file is saved

# Inspect max OOB FP
#chkObsOOB <- subset(glbObsOOB, !label.fctr.All.X..rcv.glmnet.is.acc)
#chkObsOOBFP <- subset(chkObsOOB, label.fctr.All.X..rcv.glmnet == "left_eye_center") %>% dplyr::mutate(Probability1 = label.fctr.All.X..rcv.glmnet.prob) %>% select(-.src, -.pos, -x, -y) %>% lclgetfltout_df() %>% mutate(obj.distance = (((as.numeric(x) - left_eye_center_x.int) ^ 2) + ((as.numeric(y) - left_eye_center_y.int) ^ 2)) ^ 0.5) %>% dplyr::top_n(5, obj.distance) %>% dplyr::top_n(5, -patch.cor)
#
#newImgObs <- glbObsNew[(glbObsNew$ImageId == "Test#0001"), ]; print(newImgObs[which.max(newImgObs$label.fctr.Final..rcv.glmnet.prob), ])
#OOBImgObs <- glbObsOOB[(glbObsOOB$ImageId == "Train#0003"), ]; print(OOBImgObs[which.max(OOBImgObs$label.fctr.All.X..rcv.glmnet.prob), ])

#mygetImage(which(glbObsAll[, glbFeatsId] == "Train#0003"), names(glbFeatsImage)[1], plot = TRUE, featHighlight = c("left_eye_center_x", "left_eye_center_y"), ovrlHighlight = c(66, 35))

# Depict process
glb_analytics_pn <- petrinet(name = "glb_analytics_pn",
                        trans_df = data.frame(id = 1:6,
    name = c("data.training.all","data.new",
           "model.selected","model.final",
           "data.training.all.prediction","data.new.prediction"),
    x=c(   -5,-5,-15,-25,-25,-35),
    y=c(   -5, 5,  0,  0, -5,  5)
                        ),
                        places_df=data.frame(id=1:4,
    name=c("bgn","fit.data.training.all","predict.data.new","end"),
    x=c(   -0,   -20,                    -30,               -40),
    y=c(    0,     0,                      0,                 0),
    M0=c(   3,     0,                      0,                 0)
                        ),
                        arcs_df = data.frame(
    begin = c("bgn","bgn","bgn",        
            "data.training.all","model.selected","fit.data.training.all",
            "fit.data.training.all","model.final",    
            "data.new","predict.data.new",
            "data.training.all.prediction","data.new.prediction"),
    end   = c("data.training.all","data.new","model.selected",
            "fit.data.training.all","fit.data.training.all","model.final",
            "data.training.all.prediction","predict.data.new",
            "predict.data.new","data.new.prediction",
            "end","end")
                        ))
#print(ggplot.petrinet(glb_analytics_pn))
print(ggplot.petrinet(glb_analytics_pn) + coord_flip())
## Loading required package: grid

glb_analytics_avl_objs <- NULL

glb_chunks_df <- myadd_chunk(NULL, 
                             ifelse(is.null(glbChunks$first), "import.data", glbChunks$first))
##         label step_major step_minor label_minor   bgn end elapsed
## 1 import.data          1          0           0 5.768  NA      NA

Step 1.0: import data

chunk option: eval=

## [1] "Reading file ./data/train2016.csv..."
## [1] "dimensions of data in ./data/train2016.csv: 5,568 rows x 108 cols"
##   USER_ID  YOB Gender              Income            HouseholdStatus
## 1       1 1938   Male                               Married (w/kids)
## 2       4 1970 Female       over $150,000 Domestic Partners (w/kids)
## 3       5 1997   Male  $75,000 - $100,000           Single (no kids)
## 4       8 1983   Male $100,001 - $150,000           Married (w/kids)
## 5       9 1984 Female   $50,000 - $74,999           Married (w/kids)
## 6      10 1997 Female       over $150,000           Single (no kids)
##        EducationLevel      Party Q124742 Q124122 Q123464 Q123621 Q122769
## 1                       Democrat      No              No      No      No
## 2   Bachelor's Degree   Democrat             Yes      No      No      No
## 3 High School Diploma Republican             Yes     Yes      No        
## 4   Bachelor's Degree   Democrat      No     Yes      No     Yes      No
## 5 High School Diploma Republican      No     Yes      No      No      No
## 6        Current K-12   Democrat                                      No
##   Q122770 Q122771 Q122120 Q121699 Q121700 Q120978 Q121011 Q120379 Q120650
## 1     Yes  Public      No     Yes      No              No      No     Yes
## 2     Yes  Public      No     Yes      No     Yes      No      No     Yes
## 3     Yes Private      No      No      No     Yes      No      No     Yes
## 4      No  Public      No     Yes      No     Yes      No      No     Yes
## 5     Yes  Public      No     Yes      No     Yes     Yes      No     Yes
## 6     Yes  Public      No      No      No     Yes      No     Yes     Yes
##   Q120472     Q120194 Q120012 Q120014 Q119334 Q119851   Q119650 Q118892
## 1           Try first      No      No             Yes               Yes
## 2 Science Study first     Yes     Yes      No      No Receiving      No
## 3 Science Study first             Yes      No     Yes Receiving      No
## 4 Science   Try first      No     Yes     Yes      No    Giving     Yes
## 5     Art   Try first     Yes      No      No      No    Giving      No
## 6 Science   Try first     Yes     Yes      No     Yes Receiving      No
##   Q118117    Q118232 Q118233 Q118237     Q117186        Q117193 Q116797
## 1     Yes   Idealist      No      No                                Yes
## 2      No Pragmatist      No      No Cool headed Standard hours      No
## 3     Yes Pragmatist      No     Yes Cool headed      Odd hours      No
## 4      No   Idealist      No      No Cool headed Standard hours      No
## 5      No   Idealist     Yes     Yes  Hot headed Standard hours      No
## 6      No Pragmatist      No      No             Standard hours        
##   Q116881 Q116953 Q116601 Q116441 Q116448 Q116197 Q115602 Q115777 Q115610
## 1   Happy     Yes     Yes      No      No    P.M.     Yes   Start     Yes
## 2   Happy     Yes     Yes     Yes      No    A.M.      No     End     Yes
## 3   Right     Yes      No      No     Yes    A.M.     Yes   Start     Yes
## 4   Happy     Yes     Yes      No      No    A.M.     Yes   Start     Yes
## 5   Happy     Yes     Yes      No     Yes    P.M.      No     End      No
## 6                                                                        
##   Q115611       Q115899 Q115390 Q114961 Q114748 Q115195 Q114517    Q114386
## 1      No Circumstances     Yes     Yes     Yes     Yes      No           
## 2      No            Me     Yes     Yes      No     Yes      No Mysterious
## 3     Yes Circumstances      No     Yes      No     Yes     Yes Mysterious
## 4      No Circumstances     Yes      No      No     Yes      No        TMI
## 5      No            Me      No     Yes     Yes     Yes     Yes        TMI
## 6                                                                         
##   Q113992 Q114152 Q113583    Q113584 Q113181 Q112478 Q112512 Q112270
## 1     Yes     Yes    Talk Technology      No      No     Yes        
## 2      No      No                                                   
## 3      No      No   Tunes Technology     Yes     Yes     Yes     Yes
## 4      No      No    Talk     People      No     Yes     Yes     Yes
## 5     Yes      No   Tunes     People      No      No     Yes      No
## 6                                                                   
##   Q111848    Q111580 Q111220 Q110740 Q109367       Q108950 Q109244 Q108855
## 1      No  Demanding      No              No      Cautious      No    Yes!
## 2                                Mac     Yes      Cautious      No  Umm...
## 3      No Supportive      No      PC      No      Cautious      No  Umm...
## 4     Yes Supportive      No     Mac     Yes Risk-friendly      No  Umm...
## 5      No  Demanding     Yes      PC     Yes      Cautious      No    Yes!
## 6     Yes Supportive      No      PC                                      
##   Q108617   Q108856 Q108754   Q108342 Q108343 Q107869 Q107491 Q106993
## 1      No     Space      No In-person             Yes      No     Yes
## 2      No     Space     Yes In-person      No     Yes     Yes      No
## 3      No     Space      No In-person      No      No     Yes     Yes
## 4      No Socialize     Yes    Online      No     Yes      No     Yes
## 5      No Socialize      No    Online      No      No     Yes     Yes
## 6                           In-person      No      No     Yes     Yes
##       Q106997 Q106272 Q106388 Q106389 Q106042 Q105840 Q105655 Q104996
## 1 Yay people!     Yes      No     Yes     Yes              No     Yes
## 2 Yay people!     Yes     Yes     Yes     Yes     Yes      No     Yes
## 3 Grrr people     Yes      No      No      No      No      No      No
## 4 Grrr people      No      No     Yes     Yes      No     Yes     Yes
## 5 Yay people!     Yes      No     Yes     Yes     Yes     Yes      No
## 6 Grrr people     Yes      No     Yes     Yes      No      No     Yes
##   Q103293 Q102906 Q102674 Q102687 Q102289 Q102089   Q101162 Q101163
## 1      No      No      No     Yes      No     Own  Optimist        
## 2                                                                  
## 3     Yes      No      No     Yes      No     Own Pessimist     Mom
## 4      No      No      No     Yes     Yes     Own  Optimist     Mom
## 5      No      No     Yes      No      No     Own  Optimist     Mom
## 6     Yes     Yes      No     Yes                                  
##   Q101596 Q100689 Q100680 Q100562 Q99982 Q100010 Q99716 Q99581 Q99480
## 1     Yes     Yes      No      No   Nope     Yes     No     No       
## 2                                                                  No
## 3      No      No      No      No   Nope     Yes     No     No     No
## 4      No      No      No     Yes Check!      No     No     No    Yes
## 5      No     Yes     Yes     Yes   Nope     Yes     No     No    Yes
## 6                                                                    
##   Q98869 Q98578     Q98059 Q98078 Q98197 Q96024
## 1     No        Only-child     No     No    Yes
## 2     No     No Only-child    Yes     No     No
## 3    Yes     No        Yes     No    Yes     No
## 4    Yes     No        Yes     No     No    Yes
## 5     No     No        Yes     No     No    Yes
## 6                                              
##      USER_ID  YOB Gender              Income             HouseholdStatus
## 193      245 1964   Male       over $150,000            Married (w/kids)
## 848     1046 1953   Male $100,001 - $150,000 Domestic Partners (no kids)
## 2836    3530 1995   Male                                Single (no kids)
## 4052    5050 1945 Female  $75,000 - $100,000            Married (w/kids)
## 4093    5107 1980 Female $100,001 - $150,000            Married (w/kids)
## 5509    6888 1998 Female       under $25,000            Single (no kids)
##             EducationLevel      Party Q124742 Q124122 Q123464 Q123621
## 193      Bachelor's Degree Republican     Yes     Yes      No     Yes
## 848                          Democrat                                
## 2836 Current Undergraduate   Democrat     Yes     Yes     Yes      No
## 4052     Bachelor's Degree Republican                                
## 4093     Bachelor's Degree   Democrat                      No      No
## 5509          Current K-12 Republican                                
##      Q122769 Q122770 Q122771 Q122120 Q121699 Q121700 Q120978 Q121011
## 193       No     Yes  Public      No     Yes      No     Yes      No
## 848                                                                 
## 2836             Yes  Public     Yes      No      No     Yes     Yes
## 4052              No  Public                                        
## 4093      No      No Private      No                                
## 5509                                                     Yes     Yes
##      Q120379 Q120650 Q120472     Q120194 Q120012 Q120014 Q119334 Q119851
## 193       No     Yes Science   Try first     Yes     Yes     Yes      No
## 848                                                                     
## 2836     Yes     Yes     Art Study first      No     Yes             Yes
## 4052                                                                    
## 4093                                                         Yes        
## 5509     Yes      No     Art Study first     Yes      No     Yes      No
##      Q119650 Q118892 Q118117    Q118232 Q118233 Q118237     Q117186
## 193   Giving     Yes      No   Idealist     Yes     Yes  Hot headed
## 848                                                                
## 2836             Yes     Yes   Idealist     Yes      No Cool headed
## 4052                      No                 No      No            
## 4093              No      No Pragmatist      No     Yes            
## 5509  Giving      No                                               
##             Q117193 Q116797 Q116881 Q116953 Q116601 Q116441 Q116448
## 193  Standard hours      No   Happy     Yes     Yes      No      No
## 848                                                                
## 2836      Odd hours      No   Happy     Yes     Yes              No
## 4052                                                               
## 4093                                                               
## 5509                                                               
##      Q116197 Q115602 Q115777 Q115610 Q115611       Q115899 Q115390 Q114961
## 193     A.M.     Yes     End     Yes     Yes            Me      No      No
## 848                                                                       
## 2836             Yes     End     Yes      No Circumstances     Yes      No
## 4052    P.M.     Yes   Start     Yes      No                    No        
## 4093    P.M.     Yes   Start     Yes      No Circumstances                
## 5509                                                                      
##      Q114748 Q115195 Q114517    Q114386 Q113992 Q114152 Q113583    Q113584
## 193      Yes      No     Yes        TMI      No     Yes   Tunes Technology
## 848                                                                       
## 2836     Yes      No      No Mysterious      No     Yes   Tunes     People
## 4052      No     Yes                                                      
## 4093                                                      Tunes     People
## 5509                                                                      
##      Q113181 Q112478 Q112512 Q112270 Q111848    Q111580 Q111220 Q110740
## 193       No     Yes             Yes     Yes Supportive      No     Mac
## 848                                                                    
## 2836     Yes     Yes     Yes      No     Yes  Demanding     Yes      PC
## 4052                                                                   
## 4093                                     Yes Supportive                
## 5509                                                                   
##      Q109367       Q108950 Q109244 Q108855 Q108617   Q108856 Q108754
## 193       No      Cautious      No    Yes!      No Socialize      No
## 848      Yes Risk-friendly     Yes    Yes!      No     Space      No
## 2836     Yes      Cautious     Yes             Yes                  
## 4052                                                                
## 4093      No Risk-friendly      No    Yes!      No     Space      No
## 5509                                                                
##        Q108342 Q108343 Q107869 Q107491 Q106993     Q106997 Q106272 Q106388
## 193  In-person      No     Yes     Yes      No Yay people!     Yes     Yes
## 848  In-person     Yes                                                    
## 2836 In-person     Yes             Yes                         Yes      No
## 4052                                        No Grrr people                
## 4093 In-person     Yes     Yes     Yes     Yes Yay people!     Yes     Yes
## 5509                                                                      
##      Q106389 Q106042 Q105840 Q105655 Q104996 Q103293 Q102906 Q102674
## 193       No     Yes      No      No     Yes      No      No      No
## 848                                                                 
## 2836     Yes      No      No      No     Yes     Yes      No      No
## 4052                              No      No      No              No
## 4093      No      No      No      No     Yes      No      No     Yes
## 5509                                                                
##      Q102687 Q102289 Q102089  Q101162 Q101163 Q101596 Q100689 Q100680
## 193       No      No     Own Optimist     Dad     Yes     Yes      No
## 848                                                                  
## 2836     Yes     Yes    Rent Optimist     Dad      No     Yes     Yes
## 4052     Yes             Own                       No                
## 4093     Yes     Yes    Rent                               No     Yes
## 5509                                                                 
##      Q100562 Q99982 Q100010 Q99716 Q99581 Q99480 Q98869 Q98578 Q98059
## 193      Yes Check!      No     No     No    Yes    Yes     No    Yes
## 848                                                                  
## 2836     Yes Check!      No     No     No    Yes    Yes           Yes
## 4052                                                                 
## 4093      No   Nope     Yes     No    Yes    Yes    Yes     No    Yes
## 5509                                                                 
##      Q98078 Q98197 Q96024
## 193      No    Yes    Yes
## 848                    No
## 2836    Yes    Yes     No
## 4052                     
## 4093    Yes    Yes     No
## 5509                     
##      USER_ID  YOB Gender            Income  HouseholdStatus
## 5563    6955 1966   Male     over $150,000 Married (w/kids)
## 5564    6956   NA   Male                                   
## 5565    6957 2000 Female                                   
## 5566    6958 1969   Male     over $150,000                 
## 5567    6959 1986   Male $25,001 - $50,000 Married (w/kids)
## 5568    6960 1999   Male     under $25,000 Single (no kids)
##           EducationLevel      Party Q124742 Q124122 Q123464 Q123621
## 5563   Bachelor's Degree   Democrat                                
## 5564     Master's Degree   Democrat              No      No        
## 5565        Current K-12 Republican                                
## 5566   Bachelor's Degree   Democrat                             Yes
## 5567 High School Diploma Republican                                
## 5568        Current K-12 Republican                                
##      Q122769 Q122770 Q122771 Q122120 Q121699 Q121700 Q120978 Q121011
## 5563                              No     Yes      No     Yes     Yes
## 5564      No     Yes  Public             Yes                        
## 5565                  Public                             Yes        
## 5566                              No      No      No     Yes     Yes
## 5567                             Yes             Yes              No
## 5568                                     Yes      No      No        
##      Q120379 Q120650 Q120472   Q120194 Q120012 Q120014 Q119334 Q119851
## 5563                                                                  
## 5564                                                                  
## 5565     Yes     Yes     Art Try first      No     Yes     Yes     Yes
## 5566     Yes     Yes Science                                          
## 5567      No      No Science                No     Yes                
## 5568                                                                  
##        Q119650 Q118892 Q118117 Q118232 Q118233 Q118237 Q117186 Q117193
## 5563                                                                  
## 5564                                                                  
## 5565 Receiving                                                        
## 5566                                                                  
## 5567                                                                  
## 5568                                                                  
##      Q116797 Q116881 Q116953 Q116601 Q116441 Q116448 Q116197 Q115602
## 5563                                                                
## 5564                                                                
## 5565                                                                
## 5566                                                                
## 5567                                                                
## 5568                                                                
##      Q115777 Q115610 Q115611 Q115899 Q115390 Q114961 Q114748 Q115195
## 5563                                                                
## 5564                                                                
## 5565                                                                
## 5566                                                                
## 5567                                                                
## 5568                                                                
##      Q114517 Q114386 Q113992 Q114152 Q113583 Q113584 Q113181 Q112478
## 5563                                                                
## 5564                                                                
## 5565                                                                
## 5566                                                                
## 5567                                                                
## 5568                                                                
##      Q112512 Q112270 Q111848 Q111580 Q111220 Q110740 Q109367 Q108950
## 5563                                                                
## 5564                                                                
## 5565                                                                
## 5566                                                                
## 5567                                                                
## 5568                                                                
##      Q109244 Q108855 Q108617 Q108856 Q108754 Q108342 Q108343 Q107869
## 5563                                                                
## 5564                                                                
## 5565                                                                
## 5566                                                                
## 5567                                                                
## 5568                                                                
##      Q107491 Q106993 Q106997 Q106272 Q106388 Q106389 Q106042 Q105840
## 5563                                                                
## 5564                                                                
## 5565                                                                
## 5566                                                                
## 5567                                                                
## 5568                                                                
##      Q105655 Q104996 Q103293 Q102906 Q102674 Q102687 Q102289 Q102089
## 5563                                                                
## 5564                                                                
## 5565                                                                
## 5566                                                                
## 5567                                                                
## 5568                                                                
##      Q101162 Q101163 Q101596 Q100689 Q100680 Q100562 Q99982 Q100010 Q99716
## 5563                                                                      
## 5564                                                                      
## 5565                                                                      
## 5566                                                                      
## 5567                                                                      
## 5568                                                                      
##      Q99581 Q99480 Q98869 Q98578 Q98059 Q98078 Q98197 Q96024
## 5563                                                        
## 5564                                                        
## 5565                                                        
## 5566                                                        
## 5567                                                        
## 5568                                                        
## 'data.frame':    5568 obs. of  20 variables:
##  $ USER_ID        : int  1 4 5 8 9 10 11 12 13 15 ...
##  $ YOB            : int  1938 1970 1997 1983 1984 1997 1983 1996 NA 1981 ...
##  $ Gender         : chr  "Male" "Female" "Male" "Male" ...
##  $ Income         : chr  "" "over $150,000" "$75,000 - $100,000" "$100,001 - $150,000" ...
##  $ HouseholdStatus: chr  "Married (w/kids)" "Domestic Partners (w/kids)" "Single (no kids)" "Married (w/kids)" ...
##  $ EducationLevel : chr  "" "Bachelor's Degree" "High School Diploma" "Bachelor's Degree" ...
##  $ Party          : chr  "Democrat" "Democrat" "Republican" "Democrat" ...
##  $ Q124742        : chr  "No" "" "" "No" ...
##  $ Q124122        : chr  "" "Yes" "Yes" "Yes" ...
##  $ Q123464        : chr  "No" "No" "Yes" "No" ...
##  $ Q123621        : chr  "No" "No" "No" "Yes" ...
##  $ Q122769        : chr  "No" "No" "" "No" ...
##  $ Q122770        : chr  "Yes" "Yes" "Yes" "No" ...
##  $ Q122771        : chr  "Public" "Public" "Private" "Public" ...
##  $ Q122120        : chr  "No" "No" "No" "No" ...
##  $ Q121699        : chr  "Yes" "Yes" "No" "Yes" ...
##  $ Q121700        : chr  "No" "No" "No" "No" ...
##  $ Q120978        : chr  "" "Yes" "Yes" "Yes" ...
##  $ Q121011        : chr  "No" "No" "No" "No" ...
##  $ Q120379        : chr  "No" "No" "No" "No" ...
## NULL
## 'data.frame':    5568 obs. of  20 variables:
##  $ Q120650: chr  "Yes" "Yes" "Yes" "Yes" ...
##  $ Q118117: chr  "Yes" "No" "Yes" "No" ...
##  $ Q118233: chr  "No" "No" "No" "No" ...
##  $ Q118237: chr  "No" "No" "Yes" "No" ...
##  $ Q116441: chr  "No" "Yes" "No" "No" ...
##  $ Q116197: chr  "P.M." "A.M." "A.M." "A.M." ...
##  $ Q115611: chr  "No" "No" "Yes" "No" ...
##  $ Q115899: chr  "Circumstances" "Me" "Circumstances" "Circumstances" ...
##  $ Q115390: chr  "Yes" "Yes" "No" "Yes" ...
##  $ Q114748: chr  "Yes" "No" "No" "No" ...
##  $ Q115195: chr  "Yes" "Yes" "Yes" "Yes" ...
##  $ Q113584: chr  "Technology" "" "Technology" "People" ...
##  $ Q112478: chr  "No" "" "Yes" "Yes" ...
##  $ Q112270: chr  "" "" "Yes" "Yes" ...
##  $ Q111848: chr  "No" "" "No" "Yes" ...
##  $ Q106993: chr  "Yes" "No" "Yes" "Yes" ...
##  $ Q106388: chr  "No" "Yes" "No" "No" ...
##  $ Q105655: chr  "No" "No" "No" "Yes" ...
##  $ Q104996: chr  "Yes" "Yes" "No" "Yes" ...
##  $ Q102674: chr  "No" "" "No" "No" ...
## NULL
## 'data.frame':    5568 obs. of  21 variables:
##  $ Q102674: chr  "No" "" "No" "No" ...
##  $ Q102687: chr  "Yes" "" "Yes" "Yes" ...
##  $ Q102289: chr  "No" "" "No" "Yes" ...
##  $ Q102089: chr  "Own" "" "Own" "Own" ...
##  $ Q101162: chr  "Optimist" "" "Pessimist" "Optimist" ...
##  $ Q101163: chr  "" "" "Mom" "Mom" ...
##  $ Q101596: chr  "Yes" "" "No" "No" ...
##  $ Q100689: chr  "Yes" "" "No" "No" ...
##  $ Q100680: chr  "No" "" "No" "No" ...
##  $ Q100562: chr  "No" "" "No" "Yes" ...
##  $ Q99982 : chr  "Nope" "" "Nope" "Check!" ...
##  $ Q100010: chr  "Yes" "" "Yes" "No" ...
##  $ Q99716 : chr  "No" "" "No" "No" ...
##  $ Q99581 : chr  "No" "" "No" "No" ...
##  $ Q99480 : chr  "" "No" "No" "Yes" ...
##  $ Q98869 : chr  "No" "No" "Yes" "Yes" ...
##  $ Q98578 : chr  "" "No" "No" "No" ...
##  $ Q98059 : chr  "Only-child" "Only-child" "Yes" "Yes" ...
##  $ Q98078 : chr  "No" "Yes" "No" "No" ...
##  $ Q98197 : chr  "No" "No" "Yes" "No" ...
##  $ Q96024 : chr  "Yes" "No" "No" "Yes" ...
## NULL
## Warning in myprint_str_df(obsDf): [list output truncated]
## [1] "Reading file ./data/test2016.csv..."
## [1] "dimensions of data in ./data/test2016.csv: 1,392 rows x 107 cols"
##   USER_ID  YOB Gender             Income   HouseholdStatus
## 1       2 1985 Female  $25,001 - $50,000  Single (no kids)
## 2       3 1983   Male  $50,000 - $74,999  Married (w/kids)
## 3       6 1995   Male $75,000 - $100,000  Single (no kids)
## 4       7 1980 Female  $50,000 - $74,999  Single (no kids)
## 5      14 1980 Female                    Married (no kids)
## 6      28 1973   Male      over $150,000 Married (no kids)
##          EducationLevel Q124742 Q124122 Q123464 Q123621 Q122769 Q122770
## 1       Master's Degree             Yes      No     Yes      No      No
## 2 Current Undergraduate                      No             Yes     Yes
## 3          Current K-12                                                
## 4       Master's Degree     Yes     Yes      No     Yes     Yes     Yes
## 5 Current Undergraduate             Yes      No     Yes      No      No
## 6       Master's Degree      No     Yes      No     Yes      No      No
##   Q122771 Q122120 Q121699 Q121700 Q120978 Q121011 Q120379 Q120650 Q120472
## 1  Public      No     Yes     Yes     Yes      No     Yes     Yes Science
## 2  Public      No     Yes      No                                        
## 3                      No      No      No     Yes      No     Yes Science
## 4  Public      No     Yes      No     Yes      No     Yes     Yes Science
## 5  Public     Yes     Yes      No     Yes     Yes      No     Yes     Art
## 6  Public      No     Yes      No     Yes     Yes     Yes     Yes Science
##       Q120194 Q120012 Q120014 Q119334 Q119851 Q119650 Q118892 Q118117
## 1 Study first     Yes     Yes     Yes      No  Giving     Yes      No
## 2 Study first      No     Yes              No                        
## 3   Try first      No     Yes      No     Yes  Giving                
## 4   Try first     Yes      No      No     Yes  Giving     Yes     Yes
## 5   Try first     Yes     Yes     Yes     Yes  Giving      No      No
## 6   Try first     Yes     Yes      No      No  Giving      No     Yes
##      Q118232 Q118233 Q118237     Q117186        Q117193 Q116797 Q116881
## 1   Idealist      No     Yes Cool headed      Odd hours     Yes   Happy
## 2                                                                      
## 3                                                                      
## 4   Idealist      No      No Cool headed Standard hours      No   Happy
## 5   Idealist      No     Yes  Hot headed Standard hours     Yes   Happy
## 6 Pragmatist     Yes      No  Hot headed      Odd hours     Yes   Right
##   Q116953 Q116601 Q116441 Q116448 Q116197 Q115602 Q115777 Q115610 Q115611
## 1     Yes     Yes      No     Yes    A.M.     Yes     End     Yes      No
## 2     Yes     Yes                    P.M.                                
## 3     Yes                                                                
## 4     Yes      No      No     Yes    A.M.     Yes   Start     Yes      No
## 5     Yes     Yes     Yes      No    P.M.     Yes     End      No      No
## 6     Yes     Yes     Yes     Yes    P.M.             End     Yes     Yes
##         Q115899 Q115390 Q114961 Q114748 Q115195 Q114517 Q114386 Q113992
## 1            Me      No     Yes      No     Yes     Yes     TMI        
## 2                                            No                     Yes
## 3                   Yes      No     Yes     Yes      No     TMI      No
## 4            Me     Yes      No     Yes     Yes     Yes     TMI      No
## 5            Me      No      No      No     Yes      No     TMI      No
## 6 Circumstances      No     Yes      No     Yes      No     TMI     Yes
##   Q114152 Q113583    Q113584 Q113181 Q112478 Q112512 Q112270 Q111848
## 1      No   Tunes     People     Yes     Yes      No     Yes     Yes
## 2      No                         No                      No     Yes
## 3      No   Tunes Technology     Yes      No     Yes      No        
## 4     Yes    Talk     People      No      No     Yes      No     Yes
## 5           Tunes Technology      No     Yes     Yes             Yes
## 6      No    Talk Technology      No     Yes     Yes      No     Yes
##      Q111580 Q111220 Q110740 Q109367  Q108950 Q109244 Q108855 Q108617
## 1 Supportive      No             Yes Cautious     Yes    Yes!        
## 2                 No             Yes Cautious      No    Yes!      No
## 3                                 No               No              No
## 4 Supportive      No      PC      No Cautious     Yes    Yes!      No
## 5 Supportive     Yes     Mac     Yes Cautious      No    Yes!      No
## 6  Demanding      No      PC     Yes Cautious      No  Umm...      No
##   Q108856 Q108754   Q108342 Q108343 Q107869 Q107491 Q106993     Q106997
## 1             Yes In-person     Yes                                    
## 2   Space      No                       Yes     Yes     Yes Grrr people
## 3             Yes In-person      No      No     Yes     Yes Yay people!
## 4   Space      No    Online      No      No     Yes     Yes Yay people!
## 5   Space      No In-person      No      No     Yes      No Grrr people
## 6   Space      No In-person     Yes             Yes     Yes Grrr people
##   Q106272 Q106388 Q106389 Q106042 Q105840 Q105655 Q104996 Q103293 Q102906
## 1                                                                        
## 2     Yes      No      No     Yes      No     Yes      No      No        
## 3     Yes      No     Yes      No      No     Yes     Yes      No      No
## 4      No      No      No      No      No     Yes     Yes      No      No
## 5      No      No      No     Yes     Yes     Yes     Yes     Yes      No
## 6     Yes      No     Yes     Yes      No      No      No     Yes     Yes
##   Q102674 Q102687 Q102289 Q102089   Q101162 Q101163 Q101596 Q100689
## 1                                                                No
## 2                            Rent Pessimist     Dad                
## 3      No      No     Yes     Own  Optimist     Mom      No      No
## 4      No      No      No     Own  Optimist     Dad      No      No
## 5     Yes      No      No     Own Pessimist     Mom      No     Yes
## 6     Yes     Yes      No     Own Pessimist     Mom      No     Yes
##   Q100680 Q100562 Q99982 Q100010 Q99716 Q99581 Q99480 Q98869 Q98578 Q98059
## 1     Yes     Yes                                        Yes              
## 2             Yes                                        Yes           Yes
## 3     Yes     Yes   Nope      No     No     No    Yes    Yes     No    Yes
## 4     Yes     Yes   Nope     Yes     No     No     No    Yes     No    Yes
## 5     Yes     Yes   Nope     Yes     No     No    Yes     No     No    Yes
## 6     Yes     Yes   Nope     Yes     No     No    Yes     No     No    Yes
##   Q98078 Q98197 Q96024
## 1                     
## 2    Yes     No    Yes
## 3     No    Yes    Yes
## 4     No     No    Yes
## 5     No     No     No
## 6     No     No    Yes
##      USER_ID  YOB Gender              Income   HouseholdStatus
## 503     2555 1956   Male       over $150,000  Married (w/kids)
## 515     2616 1959   Male       over $150,000  Married (w/kids)
## 857     4346 1990 Female   $50,000 - $74,999                  
## 950     4814 1969   Male  $75,000 - $100,000  Married (w/kids)
## 1207    6057 1937 Female   $25,001 - $50,000 Married (no kids)
## 1255    6285 1976 Female $100,001 - $150,000 Married (no kids)
##         EducationLevel Q124742 Q124122 Q123464 Q123621 Q122769 Q122770
## 503  Bachelor's Degree      No      No      No     Yes      No     Yes
## 515  Bachelor's Degree                                                
## 857  Bachelor's Degree                                                
## 950  Bachelor's Degree             Yes      No     Yes      No      No
## 1207 Bachelor's Degree                                      No     Yes
## 1255 Bachelor's Degree                                                
##      Q122771 Q122120 Q121699 Q121700 Q120978 Q121011 Q120379 Q120650
## 503  Private      No     Yes      No      No     Yes      No     Yes
## 515               No      No                                        
## 857               No     Yes      No      No      No      No     Yes
## 950   Public     Yes     Yes      No     Yes     Yes      No     Yes
## 1207  Public      No     Yes      No      No      No              No
## 1255                                                                
##      Q120472     Q120194 Q120012 Q120014 Q119334 Q119851   Q119650 Q118892
## 503  Science Study first      No     Yes      No     Yes    Giving     Yes
## 515                                                                    Yes
## 857  Science Study first      No      No     Yes      No Receiving     Yes
## 950  Science Study first      No      No      No      No    Giving      No
## 1207         Study first      No      No             Yes Receiving     Yes
## 1255                                                                      
##      Q118117    Q118232 Q118233 Q118237     Q117186        Q117193 Q116797
## 503       No Pragmatist      No      No Cool headed Standard hours      No
## 515       No Pragmatist      No     Yes Cool headed Standard hours      No
## 857      Yes Pragmatist      No      No Cool headed      Odd hours      No
## 950       No Pragmatist      No     Yes  Hot headed      Odd hours     Yes
## 1207      No Pragmatist      No      No  Hot headed                     No
## 1255                                                                      
##      Q116881 Q116953 Q116601 Q116441 Q116448 Q116197 Q115602 Q115777
## 503    Happy     Yes     Yes      No      No    A.M.     Yes     End
## 515    Right     Yes     Yes      No     Yes             Yes        
## 857    Right     Yes     Yes      No      No    A.M.     Yes   Start
## 950    Happy     Yes     Yes     Yes      No    P.M.     Yes   Start
## 1207   Happy     Yes     Yes      No      No    A.M.     Yes   Start
## 1255                     Yes      No     Yes    A.M.     Yes   Start
##      Q115610 Q115611       Q115899 Q115390 Q114961 Q114748 Q115195 Q114517
## 503      Yes     Yes            Me      No      No      No     Yes     Yes
## 515      Yes      No            Me     Yes      No     Yes     Yes      No
## 857      Yes      No            Me              No      No      No     Yes
## 950      Yes      No            Me     Yes      No     Yes      No      No
## 1207      No      No Circumstances     Yes      No     Yes      No     Yes
## 1255     Yes      No Circumstances      No     Yes      No     Yes     Yes
##         Q114386 Q113992 Q114152 Q113583    Q113584 Q113181 Q112478 Q112512
## 503         TMI     Yes     Yes   Tunes     People     Yes      No     Yes
## 515                  No     Yes    Talk Technology                        
## 857  Mysterious      No      No   Tunes     People      No      No      No
## 950  Mysterious      No      No   Tunes     People     Yes     Yes     Yes
## 1207                Yes      No    Talk                                Yes
## 1255        TMI             Yes                                Yes     Yes
##      Q112270 Q111848    Q111580 Q111220 Q110740 Q109367       Q108950
## 503       No     Yes  Demanding      No      PC      No      Cautious
## 515       No     Yes                 No     Mac     Yes              
## 857      Yes     Yes Supportive      No     Mac      No Risk-friendly
## 950       No     Yes Supportive     Yes      PC      No      Cautious
## 1207                 Supportive      No      PC              Cautious
## 1255     Yes     Yes  Demanding      No     Mac                      
##      Q109244 Q108855 Q108617 Q108856 Q108754   Q108342 Q108343 Q107869
## 503       No  Umm...      No   Space      No In-person      No     Yes
## 515                                                                   
## 857      Yes  Umm...      No   Space      No In-person      No     Yes
## 950       No    Yes!      No   Space      No In-person      No      No
## 1207            Yes!      No   Space      No In-person      No     Yes
## 1255                                                                  
##      Q107491 Q106993     Q106997 Q106272 Q106388 Q106389 Q106042 Q105840
## 503      Yes     Yes Yay people!     Yes      No      No     Yes      No
## 515                                                                   No
## 857       No     Yes Grrr people     Yes      No     Yes      No      No
## 950      Yes      No Grrr people     Yes     Yes      No      No      No
## 1207     Yes     Yes                 Yes                                
## 1255                                                                    
##      Q105655 Q104996 Q103293 Q102906 Q102674 Q102687 Q102289 Q102089
## 503       No     Yes      No      No      No     Yes      No     Own
## 515      Yes     Yes                                                
## 857       No     Yes     Yes      No      No     Yes     Yes     Own
## 950      Yes     Yes     Yes      No      No     Yes      No     Own
## 1207     Yes                                                        
## 1255                                                                
##        Q101162 Q101163 Q101596 Q100689 Q100680 Q100562 Q99982 Q100010
## 503  Pessimist     Mom     Yes     Yes      No     Yes Check!     Yes
## 515                                                    Check!     Yes
## 857   Optimist     Mom      No     Yes     Yes      No   Nope     Yes
## 950  Pessimist     Mom     Yes      No      No      No Check!     Yes
## 1207                                                                 
## 1255                                                                 
##      Q99716 Q99581 Q99480 Q98869 Q98578 Q98059 Q98078 Q98197 Q96024
## 503      No     No    Yes    Yes     No    Yes    Yes    Yes    Yes
## 515             No    Yes    Yes           Yes     No    Yes    Yes
## 857      No    Yes    Yes    Yes     No    Yes     No     No     No
## 950      No     No    Yes    Yes     No    Yes     No    Yes    Yes
## 1207                                                               
## 1255                                                               
##      USER_ID  YOB Gender              Income             HouseholdStatus
## 1387    6922 1988   Male   $50,000 - $74,999            Single (no kids)
## 1388    6928 1977 Female   $50,000 - $74,999 Domestic Partners (no kids)
## 1389    6930 1998 Female $100,001 - $150,000            Single (no kids)
## 1390    6941 1989   Male   $25,001 - $50,000           Married (no kids)
## 1391    6946 1996   Male                                                
## 1392    6947   NA Female                                                
##         EducationLevel Q124742 Q124122 Q123464 Q123621 Q122769 Q122770
## 1387   Master's Degree                                                
## 1388   Master's Degree                                                
## 1389      Current K-12                                      No      No
## 1390 Bachelor's Degree                                                
## 1391      Current K-12                                                
## 1392                       Yes     Yes      No      No      No      No
##      Q122771 Q122120 Q121699 Q121700 Q120978 Q121011 Q120379 Q120650
## 1387                     Yes     Yes     Yes     Yes     Yes     Yes
## 1388                             Yes              No             Yes
## 1389  Public     Yes     Yes     Yes     Yes     Yes     Yes     Yes
## 1390             Yes     Yes      No      No      No                
## 1391             Yes      No      No     Yes      No     Yes     Yes
## 1392  Public     Yes     Yes      No     Yes     Yes     Yes     Yes
##      Q120472     Q120194 Q120012 Q120014 Q119334 Q119851 Q119650 Q118892
## 1387 Science   Try first      No     Yes     Yes      No  Giving        
## 1388     Art                                                            
## 1389     Art Study first     Yes      No     Yes      No  Giving        
## 1390                                                                    
## 1391     Art Study first     Yes     Yes     Yes      No  Giving        
## 1392     Art                  No      No      No     Yes  Giving        
##      Q118117 Q118232 Q118233 Q118237 Q117186 Q117193 Q116797 Q116881
## 1387                                                                
## 1388                                                                
## 1389                                                                
## 1390                                                                
## 1391                                                                
## 1392                                                                
##      Q116953 Q116601 Q116441 Q116448 Q116197 Q115602 Q115777 Q115610
## 1387                                                                
## 1388                                                                
## 1389                                                                
## 1390                                                                
## 1391                                                                
## 1392                                                                
##      Q115611 Q115899 Q115390 Q114961 Q114748 Q115195 Q114517 Q114386
## 1387                                                                
## 1388                                                                
## 1389                                                                
## 1390                                                                
## 1391                                                                
## 1392                                                                
##      Q113992 Q114152 Q113583 Q113584 Q113181 Q112478 Q112512 Q112270
## 1387                                                                
## 1388                                                                
## 1389                                                                
## 1390                                                                
## 1391                                                                
## 1392                                                                
##      Q111848 Q111580 Q111220 Q110740 Q109367 Q108950 Q109244 Q108855
## 1387                                                                
## 1388                                                                
## 1389                                                                
## 1390                                                                
## 1391                                                                
## 1392                                                                
##      Q108617 Q108856 Q108754 Q108342 Q108343 Q107869 Q107491 Q106993
## 1387                                                                
## 1388                                                                
## 1389                                                                
## 1390                                                                
## 1391                                                                
## 1392                                                                
##      Q106997 Q106272 Q106388 Q106389 Q106042 Q105840 Q105655 Q104996
## 1387                                                                
## 1388                                                                
## 1389                                                                
## 1390                                                                
## 1391                                                                
## 1392                                                                
##      Q103293 Q102906 Q102674 Q102687 Q102289 Q102089 Q101162 Q101163
## 1387                                                                
## 1388                                                                
## 1389                                                                
## 1390                                                                
## 1391                                                                
## 1392                                                                
##      Q101596 Q100689 Q100680 Q100562 Q99982 Q100010 Q99716 Q99581 Q99480
## 1387                                                                    
## 1388                                                                    
## 1389                                                                    
## 1390                                                                    
## 1391                                                                    
## 1392                                                                    
##      Q98869 Q98578 Q98059 Q98078 Q98197 Q96024
## 1387                                          
## 1388                                          
## 1389                                          
## 1390                                          
## 1391                                          
## 1392                                          
## 'data.frame':    1392 obs. of  20 variables:
##  $ USER_ID        : int  2 3 6 7 14 28 29 37 44 56 ...
##  $ YOB            : int  1985 1983 1995 1980 1980 1973 1968 1961 1989 1975 ...
##  $ Gender         : chr  "Female" "Male" "Male" "Female" ...
##  $ Income         : chr  "$25,001 - $50,000" "$50,000 - $74,999" "$75,000 - $100,000" "$50,000 - $74,999" ...
##  $ HouseholdStatus: chr  "Single (no kids)" "Married (w/kids)" "Single (no kids)" "Single (no kids)" ...
##  $ EducationLevel : chr  "Master's Degree" "Current Undergraduate" "Current K-12" "Master's Degree" ...
##  $ Q124742        : chr  "" "" "" "Yes" ...
##  $ Q124122        : chr  "Yes" "" "" "Yes" ...
##  $ Q123464        : chr  "No" "No" "" "No" ...
##  $ Q123621        : chr  "Yes" "" "" "Yes" ...
##  $ Q122769        : chr  "No" "Yes" "" "Yes" ...
##  $ Q122770        : chr  "No" "Yes" "" "Yes" ...
##  $ Q122771        : chr  "Public" "Public" "" "Public" ...
##  $ Q122120        : chr  "No" "No" "" "No" ...
##  $ Q121699        : chr  "Yes" "Yes" "No" "Yes" ...
##  $ Q121700        : chr  "Yes" "No" "No" "No" ...
##  $ Q120978        : chr  "Yes" "" "No" "Yes" ...
##  $ Q121011        : chr  "No" "" "Yes" "No" ...
##  $ Q120379        : chr  "Yes" "" "No" "Yes" ...
##  $ Q120650        : chr  "Yes" "" "Yes" "Yes" ...
## NULL
## 'data.frame':    1392 obs. of  20 variables:
##  $ Q120012: chr  "Yes" "No" "No" "Yes" ...
##  $ Q120014: chr  "Yes" "Yes" "Yes" "No" ...
##  $ Q118117: chr  "No" "" "" "Yes" ...
##  $ Q118237: chr  "Yes" "" "" "No" ...
##  $ Q116953: chr  "Yes" "Yes" "Yes" "Yes" ...
##  $ Q116601: chr  "Yes" "Yes" "" "No" ...
##  $ Q116448: chr  "Yes" "" "" "Yes" ...
##  $ Q116197: chr  "A.M." "P.M." "" "A.M." ...
##  $ Q115899: chr  "Me" "" "" "Me" ...
##  $ Q114961: chr  "Yes" "" "No" "No" ...
##  $ Q113584: chr  "People" "" "Technology" "People" ...
##  $ Q113181: chr  "Yes" "No" "Yes" "No" ...
##  $ Q112512: chr  "No" "" "Yes" "Yes" ...
##  $ Q108950: chr  "Cautious" "Cautious" "" "Cautious" ...
##  $ Q108617: chr  "" "No" "No" "No" ...
##  $ Q108342: chr  "In-person" "" "In-person" "Online" ...
##  $ Q107491: chr  "" "Yes" "Yes" "Yes" ...
##  $ Q106272: chr  "" "Yes" "Yes" "No" ...
##  $ Q106389: chr  "" "No" "Yes" "No" ...
##  $ Q104996: chr  "" "No" "Yes" "Yes" ...
## NULL
## 'data.frame':    1392 obs. of  21 variables:
##  $ Q102674: chr  "" "" "No" "No" ...
##  $ Q102687: chr  "" "" "No" "No" ...
##  $ Q102289: chr  "" "" "Yes" "No" ...
##  $ Q102089: chr  "" "Rent" "Own" "Own" ...
##  $ Q101162: chr  "" "Pessimist" "Optimist" "Optimist" ...
##  $ Q101163: chr  "" "Dad" "Mom" "Dad" ...
##  $ Q101596: chr  "" "" "No" "No" ...
##  $ Q100689: chr  "No" "" "No" "No" ...
##  $ Q100680: chr  "Yes" "" "Yes" "Yes" ...
##  $ Q100562: chr  "Yes" "Yes" "Yes" "Yes" ...
##  $ Q99982 : chr  "" "" "Nope" "Nope" ...
##  $ Q100010: chr  "" "" "No" "Yes" ...
##  $ Q99716 : chr  "" "" "No" "No" ...
##  $ Q99581 : chr  "" "" "No" "No" ...
##  $ Q99480 : chr  "" "" "Yes" "No" ...
##  $ Q98869 : chr  "Yes" "Yes" "Yes" "Yes" ...
##  $ Q98578 : chr  "" "" "No" "No" ...
##  $ Q98059 : chr  "" "Yes" "Yes" "Yes" ...
##  $ Q98078 : chr  "" "Yes" "No" "No" ...
##  $ Q98197 : chr  "" "No" "Yes" "No" ...
##  $ Q96024 : chr  "" "Yes" "Yes" "Yes" ...
## NULL
## Warning in myprint_str_df(obsDf): [list output truncated]
## [1] "Creating new feature: .pos..."
## [1] "Creating new feature: YOB.Age.fctr..."
## [1] "Creating new feature: YOB.Age.dff..."
## [1] "Creating new feature: Gender.fctr..."
## [1] "Creating new feature: Income.fctr..."
## [1] "Creating new feature: Hhold.fctr..."
## [1] "Creating new feature: Edn.fctr..."
## [1] "Creating new feature: Q124742.fctr..."
## [1] "Creating new feature: Q124122.fctr..."
## [1] "Creating new feature: Q123621.fctr..."
## [1] "Creating new feature: Q123464.fctr..."
## [1] "Creating new feature: Q122771.fctr..."
## [1] "Creating new feature: Q122770.fctr..."
## [1] "Creating new feature: Q122769.fctr..."
## [1] "Creating new feature: Q122120.fctr..."
## [1] "Creating new feature: Q121700.fctr..."
## [1] "Creating new feature: Q121699.fctr..."
## [1] "Creating new feature: Q121011.fctr..."
## [1] "Creating new feature: Q120978.fctr..."
## [1] "Creating new feature: Q120650.fctr..."
## [1] "Creating new feature: Q120472.fctr..."
## [1] "Creating new feature: Q120379.fctr..."
## [1] "Creating new feature: Q120194.fctr..."
## [1] "Creating new feature: Q120014.fctr..."
## [1] "Creating new feature: Q120012.fctr..."
## [1] "Creating new feature: Q119851.fctr..."
## [1] "Creating new feature: Q119650.fctr..."
## [1] "Creating new feature: Q119334.fctr..."
## [1] "Creating new feature: Q118892.fctr..."
## [1] "Creating new feature: Q118237.fctr..."
## [1] "Creating new feature: Q118233.fctr..."
## [1] "Creating new feature: Q118232.fctr..."
## [1] "Creating new feature: Q118117.fctr..."
## [1] "Creating new feature: Q117193.fctr..."
## [1] "Creating new feature: Q117186.fctr..."
## [1] "Creating new feature: Q116797.fctr..."
## [1] "Creating new feature: Q116881.fctr..."
## [1] "Creating new feature: Q116953.fctr..."
## [1] "Creating new feature: Q116601.fctr..."
## [1] "Creating new feature: Q116441.fctr..."
## [1] "Creating new feature: Q116448.fctr..."
## [1] "Creating new feature: Q116197.fctr..."
## [1] "Creating new feature: Q115602.fctr..."
## [1] "Creating new feature: Q115777.fctr..."
## [1] "Creating new feature: Q115610.fctr..."
## [1] "Creating new feature: Q115611.fctr..."
## [1] "Creating new feature: Q115899.fctr..."
## [1] "Creating new feature: Q115390.fctr..."
## [1] "Creating new feature: Q115195.fctr..."
## [1] "Creating new feature: Q114961.fctr..."
## [1] "Creating new feature: Q114748.fctr..."
## [1] "Creating new feature: Q114517.fctr..."
## [1] "Creating new feature: Q114386.fctr..."
## [1] "Creating new feature: Q114152.fctr..."
## [1] "Creating new feature: Q113992.fctr..."
## [1] "Creating new feature: Q113583.fctr..."
## [1] "Creating new feature: Q113584.fctr..."
## [1] "Creating new feature: Q113181.fctr..."
## [1] "Creating new feature: Q112478.fctr..."
## [1] "Creating new feature: Q112512.fctr..."
## [1] "Creating new feature: Q112270.fctr..."
## [1] "Creating new feature: Q111848.fctr..."
## [1] "Creating new feature: Q111580.fctr..."
## [1] "Creating new feature: Q111220.fctr..."
## [1] "Creating new feature: Q110740.fctr..."
## [1] "Creating new feature: Q109367.fctr..."
## [1] "Creating new feature: Q109244.fctr..."
## [1] "Creating new feature: Q108950.fctr..."
## [1] "Creating new feature: Q108855.fctr..."
## [1] "Creating new feature: Q108617.fctr..."
## [1] "Creating new feature: Q108856.fctr..."
## [1] "Creating new feature: Q108754.fctr..."
## [1] "Creating new feature: Q108342.fctr..."
## [1] "Creating new feature: Q108343.fctr..."
## [1] "Creating new feature: Q107869.fctr..."
## [1] "Creating new feature: Q107491.fctr..."
## [1] "Creating new feature: Q106993.fctr..."
## [1] "Creating new feature: Q106997.fctr..."
## [1] "Creating new feature: Q106272.fctr..."
## [1] "Creating new feature: Q106388.fctr..."
## [1] "Creating new feature: Q106389.fctr..."
## [1] "Creating new feature: Q106042.fctr..."
## [1] "Creating new feature: Q105840.fctr..."
## [1] "Creating new feature: Q105655.fctr..."
## [1] "Creating new feature: Q104996.fctr..."
## [1] "Creating new feature: Q103293.fctr..."
## [1] "Creating new feature: Q102906.fctr..."
## [1] "Creating new feature: Q102674.fctr..."
## [1] "Creating new feature: Q102687.fctr..."
## [1] "Creating new feature: Q102289.fctr..."
## [1] "Creating new feature: Q102089.fctr..."
## [1] "Creating new feature: Q101162.fctr..."
## [1] "Creating new feature: Q101163.fctr..."
## [1] "Creating new feature: Q101596.fctr..."
## [1] "Creating new feature: Q100689.fctr..."
## [1] "Creating new feature: Q100680.fctr..."
## [1] "Creating new feature: Q100562.fctr..."
## [1] "Creating new feature: Q100010.fctr..."
## [1] "Creating new feature: Q99982.fctr..."
## [1] "Creating new feature: Q99716.fctr..."
## [1] "Creating new feature: Q99581.fctr..."
## [1] "Creating new feature: Q99480.fctr..."
## [1] "Creating new feature: Q98869.fctr..."
## [1] "Creating new feature: Q98578.fctr..."
## [1] "Creating new feature: Q98197.fctr..."
## [1] "Creating new feature: Q98059.fctr..."
## [1] "Creating new feature: Q98078.fctr..."
## [1] "Creating new feature: Q96024.fctr..."
## [1] "Partition stats:"
## Loading required package: sqldf
## Loading required package: gsubfn
## Loading required package: proto
## Loading required package: RSQLite
## Loading required package: DBI
## Loading required package: tcltk
##        Party  .src   .n
## 1   Democrat Train 2951
## 2 Republican Train 2617
## 3       <NA>  Test 1392
##        Party  .src   .n
## 1   Democrat Train 2951
## 2 Republican Train 2617
## 3       <NA>  Test 1392
## Loading required package: RColorBrewer

##    .src   .n
## 1 Train 5568
## 2  Test 1392
## [1] "Running glbObsDropCondition filter: (is.na(glbObsAll[, \"Q109244\"]) | (glbObsAll[, \"Q109244\"] != \"Yes\"))"
## [1] "Partition stats:"
##        Party  .src  .n
## 1   Democrat Train 742
## 2       <NA>  Test 223
## 3 Republican Train 183
##        Party  .src  .n
## 1   Democrat Train 742
## 2       <NA>  Test 223
## 3 Republican Train 183

##    .src  .n
## 1 Train 925
## 2  Test 223
## Loading required package: lazyeval
## Loading required package: gdata
## gdata: read.xls support for 'XLS' (Excel 97-2004) files ENABLED.
## 
## gdata: read.xls support for 'XLSX' (Excel 2007+) files ENABLED.
## 
## Attaching package: 'gdata'
## The following objects are masked from 'package:dplyr':
## 
##     combine, first, last
## The following object is masked from 'package:stats':
## 
##     nobs
## The following object is masked from 'package:utils':
## 
##     object.size
## [1] "Found 0 duplicates by all features:"
## NULL
##          label step_major step_minor label_minor    bgn    end elapsed
## 1  import.data          1          0           0  5.768 11.999   6.231
## 2 inspect.data          2          0           0 11.999     NA      NA

Step 2.0: inspect data

## Warning: Removed 223 rows containing non-finite values (stat_count).
## Loading required package: reshape2

##       Party.Democrat Party.Republican Party.NA
## Test              NA               NA      223
## Train            742              183       NA
##       Party.Democrat Party.Republican Party.NA
## Test              NA               NA        1
## Train      0.8021622        0.1978378       NA
## [1] "numeric data missing in : "
## YOB 
##  48 
## [1] "numeric data w/ 0s in : "
## YOB.Age.dff 
##          49 
## [1] "numeric data w/ Infs in : "
## named integer(0)
## [1] "numeric data w/ NaNs in : "
## named integer(0)
## [1] "string data missing in : "
##          Gender          Income HouseholdStatus  EducationLevel 
##               9             163              59             136 
##           Party         Q124742         Q124122         Q123464 
##              NA             589             337             309 
##         Q123621         Q122769         Q122770         Q122771 
##             312             279             245             247 
##         Q122120         Q121699         Q121700         Q120978 
##             248             224             222             257 
##         Q121011         Q120379         Q120650         Q120472 
##             238             269             261             280 
##         Q120194         Q120012         Q120014         Q119334 
##             292             265             289             257 
##         Q119851         Q119650         Q118892         Q118117 
##             245             274             227             228 
##         Q118232         Q118233         Q118237         Q117186 
##             311             268             264             283 
##         Q117193         Q116797         Q116881         Q116953 
##             276             242             276             279 
##         Q116601         Q116441         Q116448         Q116197 
##             225             239             243             248 
##         Q115602         Q115777         Q115610         Q115611 
##             236             264             230             202 
##         Q115899         Q115390         Q114961         Q114748 
##             260             279             244             207 
##         Q115195         Q114517         Q114386         Q113992 
##             242             216             214             206 
##         Q114152         Q113583         Q113584         Q113181 
##             261             235             241             217 
##         Q112478         Q112512         Q112270         Q111848 
##             229             212             249             179 
##         Q111580         Q111220         Q110740         Q109367 
##             219             191             173              73 
##         Q108950         Q109244         Q108855         Q108617 
##              91               0             191             143 
##         Q108856         Q108754         Q108342         Q108343 
##             183             148             157             162 
##         Q107869         Q107491         Q106993         Q106997 
##             196             176             187             193 
##         Q106272         Q106388         Q106389         Q106042 
##             212             228             236             218 
##         Q105840         Q105655         Q104996         Q103293 
##             222             177             201             201 
##         Q102906         Q102674         Q102687         Q102289 
##             231             237             199             227 
##         Q102089         Q101162         Q101163         Q101596 
##             222             224             271             243 
##         Q100689         Q100680         Q100562          Q99982 
##             185             216             226             243 
##         Q100010          Q99716          Q99581          Q99480 
##             210             219             214             206 
##          Q98869          Q98578          Q98059          Q98078 
##             254             252             195             251 
##          Q98197          Q96024 
##             224             243
##        Party Party.fctr  .n
## 1   Democrat          D 742
## 2       <NA>       <NA> 223
## 3 Republican          R 183
## Warning: Removed 1 rows containing missing values (position_stack).

##       Party.fctr.D Party.fctr.R Party.fctr.NA
## Test            NA           NA           223
## Train          742          183            NA
##       Party.fctr.D Party.fctr.R Party.fctr.NA
## Test            NA           NA             1
## Train    0.8021622    0.1978378            NA

## [1] "elapsed Time (secs): 2.226000"
## Scale for 'y' is already present. Adding another scale for 'y', which
## will replace the existing scale.
## Scale for 'y' is already present. Adding another scale for 'y', which
## will replace the existing scale.

## [1] "elapsed Time (secs): 129.390000"
## [1] "elapsed Time (secs): 129.390000"
##          label step_major step_minor label_minor     bgn     end elapsed
## 2 inspect.data          2          0           0  11.999 145.665 133.666
## 3   scrub.data          2          1           1 145.666      NA      NA

Step 2.1: scrub data

```{r scrub.data, cache=FALSE, echo=FALSE, eval=myevlChunk(glbChunks, glbOut$pfx)}

## [1] "numeric data missing in : "
##        YOB Party.fctr 
##         48        223 
## [1] "numeric data w/ 0s in : "
## YOB.Age.dff 
##          49 
## [1] "numeric data w/ Infs in : "
## named integer(0)
## [1] "numeric data w/ NaNs in : "
## named integer(0)
## [1] "string data missing in : "
##          Gender          Income HouseholdStatus  EducationLevel 
##               9             163              59             136 
##           Party         Q124742         Q124122         Q123464 
##              NA             589             337             309 
##         Q123621         Q122769         Q122770         Q122771 
##             312             279             245             247 
##         Q122120         Q121699         Q121700         Q120978 
##             248             224             222             257 
##         Q121011         Q120379         Q120650         Q120472 
##             238             269             261             280 
##         Q120194         Q120012         Q120014         Q119334 
##             292             265             289             257 
##         Q119851         Q119650         Q118892         Q118117 
##             245             274             227             228 
##         Q118232         Q118233         Q118237         Q117186 
##             311             268             264             283 
##         Q117193         Q116797         Q116881         Q116953 
##             276             242             276             279 
##         Q116601         Q116441         Q116448         Q116197 
##             225             239             243             248 
##         Q115602         Q115777         Q115610         Q115611 
##             236             264             230             202 
##         Q115899         Q115390         Q114961         Q114748 
##             260             279             244             207 
##         Q115195         Q114517         Q114386         Q113992 
##             242             216             214             206 
##         Q114152         Q113583         Q113584         Q113181 
##             261             235             241             217 
##         Q112478         Q112512         Q112270         Q111848 
##             229             212             249             179 
##         Q111580         Q111220         Q110740         Q109367 
##             219             191             173              73 
##         Q108950         Q109244         Q108855         Q108617 
##              91               0             191             143 
##         Q108856         Q108754         Q108342         Q108343 
##             183             148             157             162 
##         Q107869         Q107491         Q106993         Q106997 
##             196             176             187             193 
##         Q106272         Q106388         Q106389         Q106042 
##             212             228             236             218 
##         Q105840         Q105655         Q104996         Q103293 
##             222             177             201             201 
##         Q102906         Q102674         Q102687         Q102289 
##             231             237             199             227 
##         Q102089         Q101162         Q101163         Q101596 
##             222             224             271             243 
##         Q100689         Q100680         Q100562          Q99982 
##             185             216             226             243 
##         Q100010          Q99716          Q99581          Q99480 
##             210             219             214             206 
##          Q98869          Q98578          Q98059          Q98078 
##             254             252             195             251 
##          Q98197          Q96024 
##             224             243
##            label step_major step_minor label_minor     bgn    end elapsed
## 3     scrub.data          2          1           1 145.666 180.63  34.964
## 4 transform.data          2          2           2 180.630     NA      NA

Step 2.2: transform data

##              label step_major step_minor label_minor     bgn     end
## 4   transform.data          2          2           2 180.630 180.671
## 5 extract.features          3          0           0 180.671      NA
##   elapsed
## 4   0.041
## 5      NA

Step 3.0: extract features

##                       label step_major step_minor label_minor     bgn
## 5          extract.features          3          0           0 180.671
## 6 extract.features.datetime          3          1           1 180.692
##       end elapsed
## 5 180.691    0.02
## 6      NA      NA

Step 3.1: extract features datetime

##                           label step_major step_minor label_minor     bgn
## 1 extract.features.datetime.bgn          1          0           0 180.719
##   end elapsed
## 1  NA      NA
##                       label step_major step_minor label_minor     bgn
## 6 extract.features.datetime          3          1           1 180.692
## 7    extract.features.image          3          2           2 180.732
##       end elapsed
## 6 180.731   0.039
## 7      NA      NA

Step 3.2: extract features image

```{r extract.features.image, cache=FALSE, echo=FALSE, fig.height=5, fig.width=5, eval=myevlChunk(glbChunks, glbOut$pfx)}

##                        label step_major step_minor label_minor     bgn end
## 1 extract.features.image.bgn          1          0           0 180.762  NA
##   elapsed
## 1      NA
##                        label step_major step_minor label_minor     bgn
## 1 extract.features.image.bgn          1          0           0 180.762
## 2 extract.features.image.end          2          0           0 180.772
##       end elapsed
## 1 180.772    0.01
## 2      NA      NA
##                        label step_major step_minor label_minor     bgn
## 1 extract.features.image.bgn          1          0           0 180.762
## 2 extract.features.image.end          2          0           0 180.772
##       end elapsed
## 1 180.772    0.01
## 2      NA      NA
##                    label step_major step_minor label_minor     bgn     end
## 7 extract.features.image          3          2           2 180.732 180.782
## 8 extract.features.price          3          3           3 180.783      NA
##   elapsed
## 7   0.051
## 8      NA

Step 3.3: extract features price

##                        label step_major step_minor label_minor     bgn end
## 1 extract.features.price.bgn          1          0           0 180.809  NA
##   elapsed
## 1      NA
##                    label step_major step_minor label_minor     bgn     end
## 8 extract.features.price          3          3           3 180.783 180.817
## 9  extract.features.text          3          4           4 180.818      NA
##   elapsed
## 8   0.034
## 9      NA

Step 3.4: extract features text

##                       label step_major step_minor label_minor     bgn end
## 1 extract.features.text.bgn          1          0           0 180.861  NA
##   elapsed
## 1      NA
## Warning in rm(tmp_allobs_df): object 'tmp_allobs_df' not found
## Warning in rm(tmp_trnobs_df): object 'tmp_trnobs_df' not found
##                      label step_major step_minor label_minor     bgn
## 9    extract.features.text          3          4           4 180.818
## 10 extract.features.string          3          5           5 180.875
##        end elapsed
## 9  180.875   0.057
## 10      NA      NA

Step 3.5: extract features string

##                         label step_major step_minor label_minor     bgn
## 1 extract.features.string.bgn          1          0           0 180.911
##   end elapsed
## 1  NA      NA
##                                       label step_major step_minor
## 1               extract.features.string.bgn          1          0
## 2 extract.features.stringfactorize.str.vars          2          0
##   label_minor     bgn     end elapsed
## 1           0 180.911 180.921    0.01
## 2           0 180.921      NA      NA
##            Gender            Income   HouseholdStatus    EducationLevel 
##          "Gender"          "Income" "HouseholdStatus"  "EducationLevel" 
##             Party           Q124742           Q124122           Q123464 
##           "Party"         "Q124742"         "Q124122"         "Q123464" 
##           Q123621           Q122769           Q122770           Q122771 
##         "Q123621"         "Q122769"         "Q122770"         "Q122771" 
##           Q122120           Q121699           Q121700           Q120978 
##         "Q122120"         "Q121699"         "Q121700"         "Q120978" 
##           Q121011           Q120379           Q120650           Q120472 
##         "Q121011"         "Q120379"         "Q120650"         "Q120472" 
##           Q120194           Q120012           Q120014           Q119334 
##         "Q120194"         "Q120012"         "Q120014"         "Q119334" 
##           Q119851           Q119650           Q118892           Q118117 
##         "Q119851"         "Q119650"         "Q118892"         "Q118117" 
##           Q118232           Q118233           Q118237           Q117186 
##         "Q118232"         "Q118233"         "Q118237"         "Q117186" 
##           Q117193           Q116797           Q116881           Q116953 
##         "Q117193"         "Q116797"         "Q116881"         "Q116953" 
##           Q116601           Q116441           Q116448           Q116197 
##         "Q116601"         "Q116441"         "Q116448"         "Q116197" 
##           Q115602           Q115777           Q115610           Q115611 
##         "Q115602"         "Q115777"         "Q115610"         "Q115611" 
##           Q115899           Q115390           Q114961           Q114748 
##         "Q115899"         "Q115390"         "Q114961"         "Q114748" 
##           Q115195           Q114517           Q114386           Q113992 
##         "Q115195"         "Q114517"         "Q114386"         "Q113992" 
##           Q114152           Q113583           Q113584           Q113181 
##         "Q114152"         "Q113583"         "Q113584"         "Q113181" 
##           Q112478           Q112512           Q112270           Q111848 
##         "Q112478"         "Q112512"         "Q112270"         "Q111848" 
##           Q111580           Q111220           Q110740           Q109367 
##         "Q111580"         "Q111220"         "Q110740"         "Q109367" 
##           Q108950           Q109244           Q108855           Q108617 
##         "Q108950"         "Q109244"         "Q108855"         "Q108617" 
##           Q108856           Q108754           Q108342           Q108343 
##         "Q108856"         "Q108754"         "Q108342"         "Q108343" 
##           Q107869           Q107491           Q106993           Q106997 
##         "Q107869"         "Q107491"         "Q106993"         "Q106997" 
##           Q106272           Q106388           Q106389           Q106042 
##         "Q106272"         "Q106388"         "Q106389"         "Q106042" 
##           Q105840           Q105655           Q104996           Q103293 
##         "Q105840"         "Q105655"         "Q104996"         "Q103293" 
##           Q102906           Q102674           Q102687           Q102289 
##         "Q102906"         "Q102674"         "Q102687"         "Q102289" 
##           Q102089           Q101162           Q101163           Q101596 
##         "Q102089"         "Q101162"         "Q101163"         "Q101596" 
##           Q100689           Q100680           Q100562            Q99982 
##         "Q100689"         "Q100680"         "Q100562"          "Q99982" 
##           Q100010            Q99716            Q99581            Q99480 
##         "Q100010"          "Q99716"          "Q99581"          "Q99480" 
##            Q98869            Q98578            Q98059            Q98078 
##          "Q98869"          "Q98578"          "Q98059"          "Q98078" 
##            Q98197            Q96024              .src 
##          "Q98197"          "Q96024"            ".src"
##                      label step_major step_minor label_minor     bgn
## 10 extract.features.string          3          5           5 180.875
## 11    extract.features.end          3          6           6 180.944
##        end elapsed
## 10 180.943   0.069
## 11      NA      NA

Step 3.6: extract features end

## time trans    "bgn " "fit.data.training.all " "predict.data.new " "end " 
## 0.0000   multiple enabled transitions:  data.training.all data.new model.selected    firing:  data.training.all 
## 1.0000    1   2 1 0 0 
## 1.0000   multiple enabled transitions:  data.training.all data.new model.selected model.final data.training.all.prediction   firing:  data.new 
## 2.0000    2   1 1 1 0

##                   label step_major step_minor label_minor     bgn     end
## 11 extract.features.end          3          6           6 180.944 181.818
## 12  manage.missing.data          4          0           0 181.818      NA
##    elapsed
## 11   0.874
## 12      NA

Step 4.0: manage missing data

## [1] "numeric data missing in : "
##        YOB Party.fctr 
##         48        223 
## [1] "numeric data w/ 0s in : "
## YOB.Age.dff 
##          49 
## [1] "numeric data w/ Infs in : "
## named integer(0)
## [1] "numeric data w/ NaNs in : "
## named integer(0)
## [1] "string data missing in : "
##          Gender          Income HouseholdStatus  EducationLevel 
##               9             163              59             136 
##           Party         Q124742         Q124122         Q123464 
##              NA             589             337             309 
##         Q123621         Q122769         Q122770         Q122771 
##             312             279             245             247 
##         Q122120         Q121699         Q121700         Q120978 
##             248             224             222             257 
##         Q121011         Q120379         Q120650         Q120472 
##             238             269             261             280 
##         Q120194         Q120012         Q120014         Q119334 
##             292             265             289             257 
##         Q119851         Q119650         Q118892         Q118117 
##             245             274             227             228 
##         Q118232         Q118233         Q118237         Q117186 
##             311             268             264             283 
##         Q117193         Q116797         Q116881         Q116953 
##             276             242             276             279 
##         Q116601         Q116441         Q116448         Q116197 
##             225             239             243             248 
##         Q115602         Q115777         Q115610         Q115611 
##             236             264             230             202 
##         Q115899         Q115390         Q114961         Q114748 
##             260             279             244             207 
##         Q115195         Q114517         Q114386         Q113992 
##             242             216             214             206 
##         Q114152         Q113583         Q113584         Q113181 
##             261             235             241             217 
##         Q112478         Q112512         Q112270         Q111848 
##             229             212             249             179 
##         Q111580         Q111220         Q110740         Q109367 
##             219             191             173              73 
##         Q108950         Q109244         Q108855         Q108617 
##              91               0             191             143 
##         Q108856         Q108754         Q108342         Q108343 
##             183             148             157             162 
##         Q107869         Q107491         Q106993         Q106997 
##             196             176             187             193 
##         Q106272         Q106388         Q106389         Q106042 
##             212             228             236             218 
##         Q105840         Q105655         Q104996         Q103293 
##             222             177             201             201 
##         Q102906         Q102674         Q102687         Q102289 
##             231             237             199             227 
##         Q102089         Q101162         Q101163         Q101596 
##             222             224             271             243 
##         Q100689         Q100680         Q100562          Q99982 
##             185             216             226             243 
##         Q100010          Q99716          Q99581          Q99480 
##             210             219             214             206 
##          Q98869          Q98578          Q98059          Q98078 
##             254             252             195             251 
##          Q98197          Q96024 
##             224             243
## [1] "numeric data missing in : "
##        YOB Party.fctr 
##         48        223 
## [1] "numeric data w/ 0s in : "
## YOB.Age.dff 
##          49 
## [1] "numeric data w/ Infs in : "
## named integer(0)
## [1] "numeric data w/ NaNs in : "
## named integer(0)
## [1] "string data missing in : "
##          Gender          Income HouseholdStatus  EducationLevel 
##               9             163              59             136 
##           Party         Q124742         Q124122         Q123464 
##              NA             589             337             309 
##         Q123621         Q122769         Q122770         Q122771 
##             312             279             245             247 
##         Q122120         Q121699         Q121700         Q120978 
##             248             224             222             257 
##         Q121011         Q120379         Q120650         Q120472 
##             238             269             261             280 
##         Q120194         Q120012         Q120014         Q119334 
##             292             265             289             257 
##         Q119851         Q119650         Q118892         Q118117 
##             245             274             227             228 
##         Q118232         Q118233         Q118237         Q117186 
##             311             268             264             283 
##         Q117193         Q116797         Q116881         Q116953 
##             276             242             276             279 
##         Q116601         Q116441         Q116448         Q116197 
##             225             239             243             248 
##         Q115602         Q115777         Q115610         Q115611 
##             236             264             230             202 
##         Q115899         Q115390         Q114961         Q114748 
##             260             279             244             207 
##         Q115195         Q114517         Q114386         Q113992 
##             242             216             214             206 
##         Q114152         Q113583         Q113584         Q113181 
##             261             235             241             217 
##         Q112478         Q112512         Q112270         Q111848 
##             229             212             249             179 
##         Q111580         Q111220         Q110740         Q109367 
##             219             191             173              73 
##         Q108950         Q109244         Q108855         Q108617 
##              91               0             191             143 
##         Q108856         Q108754         Q108342         Q108343 
##             183             148             157             162 
##         Q107869         Q107491         Q106993         Q106997 
##             196             176             187             193 
##         Q106272         Q106388         Q106389         Q106042 
##             212             228             236             218 
##         Q105840         Q105655         Q104996         Q103293 
##             222             177             201             201 
##         Q102906         Q102674         Q102687         Q102289 
##             231             237             199             227 
##         Q102089         Q101162         Q101163         Q101596 
##             222             224             271             243 
##         Q100689         Q100680         Q100562          Q99982 
##             185             216             226             243 
##         Q100010          Q99716          Q99581          Q99480 
##             210             219             214             206 
##          Q98869          Q98578          Q98059          Q98078 
##             254             252             195             251 
##          Q98197          Q96024 
##             224             243
##                  label step_major step_minor label_minor     bgn     end
## 12 manage.missing.data          4          0           0 181.818 182.334
## 13        cluster.data          5          0           0 182.334      NA
##    elapsed
## 12   0.516
## 13      NA

Step 5.0: cluster data

```{r cluster.data, cache=FALSE, echo=FALSE, eval=myevlChunk(glbChunks, glbOut$pfx)}

Step 5.0: cluster data

Step 5.0: cluster data

```{r select.features, cache=FALSE, echo=FALSE, eval=myevlChunk(glbChunks, glbOut$pfx)}

Step 5.0: cluster data

fit.models_0_chunk_df <- myadd_chunk(NULL, "fit.models_0_bgn", label.minor = "setup")
# load(paste0(glbOut$pfx, "dsk.RData"))

glbgetModelSelectFormula <- function() {
    model_evl_terms <- c(NULL)
    # min.aic.fit might not be avl
    lclMdlEvlCriteria <- 
        glbMdlMetricsEval[glbMdlMetricsEval %in% names(glb_models_df)]
    for (metric in lclMdlEvlCriteria)
        model_evl_terms <- c(model_evl_terms, 
                             ifelse(length(grep("max", metric)) > 0, "-", "+"), metric)
    if (glb_is_classification && glb_is_binomial)
        model_evl_terms <- c(model_evl_terms, "-", "opt.prob.threshold.OOB")
    model_sel_frmla <- as.formula(paste(c("~ ", model_evl_terms), collapse = " "))
    return(model_sel_frmla)
}

glbgetDisplayModelsDf <- function() {
    dsp_models_cols <- c("id", 
                    glbMdlMetricsEval[glbMdlMetricsEval %in% names(glb_models_df)],
                    grep("opt.", names(glb_models_df), fixed = TRUE, value = TRUE)) 
    dsp_models_df <- 
        #orderBy(glbgetModelSelectFormula(), glb_models_df)[, c("id", glbMdlMetricsEval)]
        orderBy(glbgetModelSelectFormula(), glb_models_df)[, dsp_models_cols]    
    nCvMdl <- sapply(glb_models_lst, function(mdl) nrow(mdl$results))
    nParams <- sapply(glb_models_lst, function(mdl) ifelse(mdl$method == "custom", 0, 
        nrow(subset(modelLookup(mdl$method), parameter != "parameter"))))
    
#     nCvMdl <- nCvMdl[names(nCvMdl) != "avNNet"]
#     nParams <- nParams[names(nParams) != "avNNet"]    
    
    if (length(cvMdlProblems <- nCvMdl[nCvMdl <= nParams]) > 0) {
        print("Cross Validation issues:")
        warning("Cross Validation issues:")        
        print(cvMdlProblems)
    }
    
    pltMdls <- setdiff(names(nCvMdl), names(cvMdlProblems))
    pltMdls <- setdiff(pltMdls, names(nParams[nParams == 0]))
    
    # length(pltMdls) == 21
    png(paste0(glbOut$pfx, "bestTune.png"), width = 480 * 2, height = 480 * 4)
    grid.newpage()
    pushViewport(viewport(layout = grid.layout(ceiling(length(pltMdls) / 2.0), 2)))
    pltIx <- 1
    for (mdlId in pltMdls) {
        print(ggplot(glb_models_lst[[mdlId]], highBestTune = TRUE) + labs(title = mdlId),   
              vp = viewport(layout.pos.row = ceiling(pltIx / 2.0), 
                            layout.pos.col = ((pltIx - 1) %% 2) + 1))  
        pltIx <- pltIx + 1
    }
    dev.off()

    if (all(row.names(dsp_models_df) != dsp_models_df$id))
        row.names(dsp_models_df) <- dsp_models_df$id
    return(dsp_models_df)
}
#glbgetDisplayModelsDf()

glb_get_predictions <- function(df, mdl_id, rsp_var, prob_threshold_def=NULL, verbose=FALSE) {
    mdl <- glb_models_lst[[mdl_id]]
    
    clmnNames <- mygetPredictIds(rsp_var, mdl_id)
    predct_var_name <- clmnNames$value        
    predct_prob_var_name <- clmnNames$prob
    predct_accurate_var_name <- clmnNames$is.acc
    predct_error_var_name <- clmnNames$err
    predct_erabs_var_name <- clmnNames$err.abs

    if (glb_is_regression) {
        df[, predct_var_name] <- predict(mdl, newdata=df, type="raw")
        if (verbose) print(myplot_scatter(df, glb_rsp_var, predct_var_name) + 
                  facet_wrap(reformulate(glbFeatsCategory), scales = "free") + 
                  stat_smooth(method="glm"))

        df[, predct_error_var_name] <- df[, predct_var_name] - df[, glb_rsp_var]
        if (verbose) print(myplot_scatter(df, predct_var_name, predct_error_var_name) + 
                  #facet_wrap(reformulate(glbFeatsCategory), scales = "free") + 
                  stat_smooth(method="auto"))
        if (verbose) print(myplot_scatter(df, glb_rsp_var, predct_error_var_name) + 
                  #facet_wrap(reformulate(glbFeatsCategory), scales = "free") + 
                  stat_smooth(method="glm"))
        
        df[, predct_erabs_var_name] <- abs(df[, predct_error_var_name])
        if (verbose) print(head(orderBy(reformulate(c("-", predct_erabs_var_name)), df)))
        
        df[, predct_accurate_var_name] <- (df[, glb_rsp_var] == df[, predct_var_name])
    }

    if (glb_is_classification && glb_is_binomial) {
        prob_threshold <- glb_models_df[glb_models_df$id == mdl_id, 
                                        "opt.prob.threshold.OOB"]
        if (is.null(prob_threshold) || is.na(prob_threshold)) {
            warning("Using default probability threshold: ", prob_threshold_def)
            if (is.null(prob_threshold <- prob_threshold_def))
                stop("Default probability threshold is NULL")
        }
        
        df[, predct_prob_var_name] <- predict(mdl, newdata = df, type = "prob")[, 2]
        df[, predct_var_name] <- 
                factor(levels(df[, glb_rsp_var])[
                    (df[, predct_prob_var_name] >=
                        prob_threshold) * 1 + 1], levels(df[, glb_rsp_var]))
    
#         if (verbose) print(myplot_scatter(df, glb_rsp_var, predct_var_name) + 
#                   facet_wrap(reformulate(glbFeatsCategory), scales = "free") + 
#                   stat_smooth(method="glm"))

        df[, predct_error_var_name] <- df[, predct_var_name] != df[, glb_rsp_var]
#         if (verbose) print(myplot_scatter(df, predct_var_name, predct_error_var_name) + 
#                   #facet_wrap(reformulate(glbFeatsCategory), scales = "free") + 
#                   stat_smooth(method="auto"))
#         if (verbose) print(myplot_scatter(df, glb_rsp_var, predct_error_var_name) + 
#                   #facet_wrap(reformulate(glbFeatsCategory), scales = "free") + 
#                   stat_smooth(method="glm"))
        
        # if prediction is a TP (true +ve), measure distance from 1.0
        tp <- which((df[, predct_var_name] == df[, glb_rsp_var]) &
                    (df[, predct_var_name] == levels(df[, glb_rsp_var])[2]))
        df[tp, predct_erabs_var_name] <- abs(1 - df[tp, predct_prob_var_name])
        #rowIx <- which.max(df[tp, predct_erabs_var_name]); df[tp, c(glbFeatsId, glb_rsp_var, predct_var_name, predct_prob_var_name, predct_erabs_var_name)][rowIx, ]
        
        # if prediction is a TN (true -ve), measure distance from 0.0
        tn <- which((df[, predct_var_name] == df[, glb_rsp_var]) &
                    (df[, predct_var_name] == levels(df[, glb_rsp_var])[1]))
        df[tn, predct_erabs_var_name] <- abs(0 - df[tn, predct_prob_var_name])
        #rowIx <- which.max(df[tn, predct_erabs_var_name]); df[tn, c(glbFeatsId, glb_rsp_var, predct_var_name, predct_prob_var_name, predct_erabs_var_name)][rowIx, ]
        
        # if prediction is a FP (flse +ve), measure distance from 0.0
        fp <- which((df[, predct_var_name] != df[, glb_rsp_var]) &
                    (df[, predct_var_name] == levels(df[, glb_rsp_var])[2]))
        df[fp, predct_erabs_var_name] <- abs(0 - df[fp, predct_prob_var_name])
        #rowIx <- which.max(df[fp, predct_erabs_var_name]); df[fp, c(glbFeatsId, glb_rsp_var, predct_var_name, predct_prob_var_name, predct_erabs_var_name)][rowIx, ]
        
        # if prediction is a FN (flse -ve), measure distance from 1.0
        fn <- which((df[, predct_var_name] != df[, glb_rsp_var]) &
                    (df[, predct_var_name] == levels(df[, glb_rsp_var])[1]))
        df[fn, predct_erabs_var_name] <- abs(1 - df[fn, predct_prob_var_name])
        #rowIx <- which.max(df[fn, predct_erabs_var_name]); df[fn, c(glbFeatsId, glb_rsp_var, predct_var_name, predct_prob_var_name, predct_erabs_var_name)][rowIx, ]

        
        if (verbose) print(head(orderBy(reformulate(c("-", predct_erabs_var_name)), df)))
        
        df[, predct_accurate_var_name] <- (df[, glb_rsp_var] == df[, predct_var_name])
    }    
    
    if (glb_is_classification && !glb_is_binomial) {
        df[, predct_var_name] <- predict(mdl, newdata = df, type = "raw")
        probCls <- predict(mdl, newdata = df, type = "prob")        
        df[, predct_prob_var_name] <- NA
        for (cls in names(probCls)) {
            mask <- (df[, predct_var_name] == cls)
            df[mask, predct_prob_var_name] <- probCls[mask, cls]
        }    
        if (verbose) print(myplot_histogram(df, predct_prob_var_name, 
                                            fill_col_name = predct_var_name))
        if (verbose) print(myplot_histogram(df, predct_prob_var_name, 
                                            facet_frmla = paste0("~", glb_rsp_var)))
        
        df[, predct_error_var_name] <- df[, predct_var_name] != df[, glb_rsp_var]
        
        # if prediction is erroneous, measure predicted class prob from actual class prob
        df[, predct_erabs_var_name] <- 0
        for (cls in names(probCls)) {
            mask <- (df[, glb_rsp_var] == cls) & (df[, predct_error_var_name])
            df[mask, predct_erabs_var_name] <- probCls[mask, cls]
        }    

        df[, predct_accurate_var_name] <- (df[, glb_rsp_var] == df[, predct_var_name])        
    }

    return(df)
}    

if (glb_is_classification && glb_is_binomial && 
        (length(unique(glbObsFit[, glb_rsp_var])) < 2))
    stop("glbObsFit$", glb_rsp_var, ": contains less than 2 unique values: ",
         paste0(unique(glbObsFit[, glb_rsp_var]), collapse=", "))

max_cor_y_x_vars <- orderBy(~ -cor.y.abs, 
        subset(glb_feats_df, (exclude.as.feat == 0) & !nzv & !is.cor.y.abs.low & 
                                is.na(cor.high.X)))[1:2, "id"]
max_cor_y_x_vars <- max_cor_y_x_vars[!is.na(max_cor_y_x_vars)]
if (length(max_cor_y_x_vars) < 2)
    max_cor_y_x_vars <- union(max_cor_y_x_vars, ".pos")

if (!is.null(glb_Baseline_mdl_var)) {
    if ((max_cor_y_x_vars[1] != glb_Baseline_mdl_var) & 
        (glb_feats_df[glb_feats_df$id == max_cor_y_x_vars[1], "cor.y.abs"] > 
         glb_feats_df[glb_feats_df$id == glb_Baseline_mdl_var, "cor.y.abs"]))
        stop(max_cor_y_x_vars[1], " has a higher correlation with ", glb_rsp_var, 
             " than the Baseline var: ", glb_Baseline_mdl_var)
}

glb_model_type <- ifelse(glb_is_regression, "regression", "classification")
    
# Model specs
# c("id.prefix", "method", "type",
#   # trainControl params
#   "preProc.method", "cv.n.folds", "cv.n.repeats", "summary.fn",
#   # train params
#   "metric", "metric.maximize", "tune.df")

# Baseline
if (!is.null(glb_Baseline_mdl_var)) {
    fit.models_0_chunk_df <- myadd_chunk(fit.models_0_chunk_df, 
                            paste0("fit.models_0_", "Baseline"), major.inc = FALSE,
                                    label.minor = "mybaseln_classfr")
    ret_lst <- myfit_mdl(mdl_id="Baseline", 
                         model_method="mybaseln_classfr",
                        indepVar=glb_Baseline_mdl_var,
                        rsp_var=glb_rsp_var,
                        fit_df=glbObsFit, OOB_df=glbObsOOB)
}    

# Most Frequent Outcome "MFO" model: mean(y) for regression
#   Not using caret's nullModel since model stats not avl
#   Cannot use rpart for multinomial classification since it predicts non-MFO
if (glb_is_classification) {
    fit.models_0_chunk_df <- myadd_chunk(fit.models_0_chunk_df, 
                                paste0("fit.models_0_", "MFO"), major.inc = FALSE,
                                        label.minor = "myMFO_classfr")

    ret_lst <- myfit_mdl(mdl_specs_lst = myinit_mdl_specs_lst(mdl_specs_lst = list(
        id.prefix = "MFO", type = glb_model_type, trainControl.method = "none",
        train.method = ifelse(glb_is_regression, "lm", "myMFO_classfr"))),
                            indepVar = ".rnorm", rsp_var = glb_rsp_var,
                            fit_df = glbObsFit, OOB_df = glbObsOOB)

        # "random" model - only for classification; 
        #   none needed for regression since it is same as MFO
    fit.models_0_chunk_df <- myadd_chunk(fit.models_0_chunk_df, 
                                paste0("fit.models_0_", "Random"), major.inc = FALSE,
                                        label.minor = "myrandom_classfr")

#stop(here"); glb2Sav(); all.equal(glb_models_df, sav_models_df)    
    ret_lst <- myfit_mdl(mdl_specs_lst = myinit_mdl_specs_lst(mdl_specs_lst = list(
        id.prefix = "Random", type = glb_model_type, trainControl.method = "none",
        train.method = "myrandom_classfr")),
                        indepVar = ".rnorm", rsp_var = glb_rsp_var,
                        fit_df = glbObsFit, OOB_df = glbObsOOB)
}

# Max.cor.Y
#   Check impact of cv
#       rpart is not a good candidate since caret does not optimize cp (only tuning parameter of rpart) well
fit.models_0_chunk_df <- myadd_chunk(fit.models_0_chunk_df, 
                        paste0("fit.models_0_", "Max.cor.Y.rcv.*X*"), major.inc = FALSE,
                                    label.minor = "glmnet")

ret_lst <- myfit_mdl(mdl_specs_lst = myinit_mdl_specs_lst(mdl_specs_lst = list(
    id.prefix = "Max.cor.Y.rcv.1X1", type = glb_model_type, trainControl.method = "none",
    train.method = "glmnet")),
                    indepVar = max_cor_y_x_vars, rsp_var = glb_rsp_var, 
                    fit_df = glbObsFit, OOB_df = glbObsOOB)

if (glbMdlCheckRcv) {
    # rcv_n_folds == 1 & rcv_n_repeats > 1 crashes
    for (rcv_n_folds in seq(3, glb_rcv_n_folds + 2, 2))
        for (rcv_n_repeats in seq(1, glb_rcv_n_repeats + 2, 2)) {
            
            # Experiment specific code to avoid caret crash
    #         lcl_tune_models_df <- rbind(data.frame()
    #                             ,data.frame(method = "glmnet", parameter = "alpha", 
    #                                         vals = "0.100 0.325 0.550 0.775 1.000")
    #                             ,data.frame(method = "glmnet", parameter = "lambda",
    #                                         vals = "9.342e-02")    
    #                                     )
            
            ret_lst <- myfit_mdl(mdl_specs_lst = myinit_mdl_specs_lst(mdl_specs_lst =
                list(
                id.prefix = paste0("Max.cor.Y.rcv.", rcv_n_folds, "X", rcv_n_repeats), 
                type = glb_model_type, 
    # tune.df = lcl_tune_models_df,            
                trainControl.method = "repeatedcv",
                trainControl.number = rcv_n_folds, 
                trainControl.repeats = rcv_n_repeats,
                trainControl.classProbs = glb_is_classification,
                trainControl.summaryFunction = glbMdlMetricSummaryFn,
                train.method = "glmnet", train.metric = glbMdlMetricSummary, 
                train.maximize = glbMdlMetricMaximize)),
                                indepVar = max_cor_y_x_vars, rsp_var = glb_rsp_var, 
                                fit_df = glbObsFit, OOB_df = glbObsOOB)
        }
    # Add parallel coordinates graph of glb_models_df[, glbMdlMetricsEval] to evaluate cv parameters
    tmp_models_cols <- c("id", "max.nTuningRuns",
                        glbMdlMetricsEval[glbMdlMetricsEval %in% names(glb_models_df)],
                        grep("opt.", names(glb_models_df), fixed = TRUE, value = TRUE)) 
    print(myplot_parcoord(obs_df = subset(glb_models_df, 
                                          grepl("Max.cor.Y.rcv.", id, fixed = TRUE), 
                                            select = -feats)[, tmp_models_cols],
                          id_var = "id"))
}
        
# Useful for stacking decisions
# fit.models_0_chunk_df <- myadd_chunk(fit.models_0_chunk_df, 
#                     paste0("fit.models_0_", "Max.cor.Y[rcv.1X1.cp.0|]"), major.inc = FALSE,
#                                     label.minor = "rpart")
# 
# ret_lst <- myfit_mdl(mdl_specs_lst = myinit_mdl_specs_lst(mdl_specs_lst = list(
#     id.prefix = "Max.cor.Y.rcv.1X1.cp.0", type = glb_model_type, trainControl.method = "none",
#     train.method = "rpart",
#     tune.df=data.frame(method="rpart", parameter="cp", min=0.0, max=0.0, by=0.1))),
#                     indepVar=max_cor_y_x_vars, rsp_var=glb_rsp_var, 
#                     fit_df=glbObsFit, OOB_df=glbObsOOB)

#stop(here"); glb2Sav(); all.equal(glb_models_df, sav_models_df)
# if (glb_is_regression || glb_is_binomial) # For multinomials this model will be run next by default
ret_lst <- myfit_mdl(mdl_specs_lst = myinit_mdl_specs_lst(mdl_specs_lst = list(
                        id.prefix = "Max.cor.Y", 
                        type = glb_model_type, trainControl.method = "repeatedcv",
                        trainControl.number = glb_rcv_n_folds, 
                        trainControl.repeats = glb_rcv_n_repeats,
                        trainControl.classProbs = glb_is_classification,
                        trainControl.summaryFunction = glbMdlMetricSummaryFn,
                        trainControl.allowParallel = glbMdlAllowParallel,                        
                        train.metric = glbMdlMetricSummary, 
                        train.maximize = glbMdlMetricMaximize,    
                        train.method = "rpart")),
                    indepVar = max_cor_y_x_vars, rsp_var = glb_rsp_var, 
                    fit_df = glbObsFit, OOB_df = glbObsOOB)

if ((length(glbFeatsDateTime) > 0) && 
    (sum(grepl(paste(names(glbFeatsDateTime), "\\.day\\.minutes\\.poly\\.", sep = ""),
               names(glbObsAll))) > 0)) {
    fit.models_0_chunk_df <- myadd_chunk(fit.models_0_chunk_df, 
                    paste0("fit.models_0_", "Max.cor.Y.Time.Poly"), major.inc = FALSE,
                                    label.minor = "glmnet")

    indepVars <- c(max_cor_y_x_vars, 
            grep(paste(names(glbFeatsDateTime), "\\.day\\.minutes\\.poly\\.", sep = ""),
                        names(glbObsAll), value = TRUE))
    indepVars <- myadjustInteractionFeats(glb_feats_df, indepVars)
    ret_lst <- myfit_mdl(mdl_specs_lst = myinit_mdl_specs_lst(mdl_specs_lst = list(
            id.prefix = "Max.cor.Y.Time.Poly", 
            type = glb_model_type, trainControl.method = "repeatedcv",
            trainControl.number = glb_rcv_n_folds, trainControl.repeats = glb_rcv_n_repeats,
            trainControl.classProbs = glb_is_classification,
            trainControl.summaryFunction = glbMdlMetricSummaryFn,
            trainControl.allowParallel = glbMdlAllowParallel,            
            train.metric = glbMdlMetricSummary, 
            train.maximize = glbMdlMetricMaximize,    
            train.method = "glmnet")),
        indepVar = indepVars,
        rsp_var = glb_rsp_var, 
        fit_df = glbObsFit, OOB_df = glbObsOOB)
}

if ((length(glbFeatsDateTime) > 0) && 
    (sum(grepl(paste(names(glbFeatsDateTime), "\\.last[[:digit:]]", sep = ""),
               names(glbObsAll))) > 0)) {
    fit.models_0_chunk_df <- myadd_chunk(fit.models_0_chunk_df, 
                    paste0("fit.models_0_", "Max.cor.Y.Time.Lag"), major.inc = FALSE,
                                    label.minor = "glmnet")

    indepVars <- c(max_cor_y_x_vars, 
            grep(paste(names(glbFeatsDateTime), "\\.last[[:digit:]]", sep = ""),
                        names(glbObsAll), value = TRUE))
    indepVars <- myadjustInteractionFeats(glb_feats_df, indepVars)
    ret_lst <- myfit_mdl(mdl_specs_lst = myinit_mdl_specs_lst(mdl_specs_lst = list(
        id.prefix = "Max.cor.Y.Time.Lag", 
        type = glb_model_type, 
        tune.df = glbMdlTuneParams,        
        trainControl.method = "repeatedcv",
        trainControl.number = glb_rcv_n_folds, trainControl.repeats = glb_rcv_n_repeats,
        trainControl.classProbs = glb_is_classification,
        trainControl.summaryFunction = glbMdlMetricSummaryFn,
        trainControl.allowParallel = glbMdlAllowParallel,        
        train.metric = glbMdlMetricSummary, 
        train.maximize = glbMdlMetricMaximize,    
        train.method = "glmnet")),
        indepVar = indepVars,
        rsp_var = glb_rsp_var, 
        fit_df = glbObsFit, OOB_df = glbObsOOB)
}

if (length(glbFeatsText) > 0) {
    fit.models_0_chunk_df <- myadd_chunk(fit.models_0_chunk_df, 
                    paste0("fit.models_0_", "Txt.*"), major.inc = FALSE,
                                    label.minor = "glmnet")

    indepVars <- c(max_cor_y_x_vars)
    for (txtFeat in names(glbFeatsText))
        indepVars <- union(indepVars, 
            grep(paste(str_to_upper(substr(txtFeat, 1, 1)), "\\.(?!([T|P]\\.))", sep = ""),
                        names(glbObsAll), perl = TRUE, value = TRUE))
    indepVars <- myadjustInteractionFeats(glb_feats_df, indepVars)
    ret_lst <- myfit_mdl(mdl_specs_lst = myinit_mdl_specs_lst(mdl_specs_lst = list(
        id.prefix = "Max.cor.Y.Text.nonTP", 
        type = glb_model_type, 
        tune.df = glbMdlTuneParams,        
        trainControl.method = "repeatedcv",
        trainControl.number = glb_rcv_n_folds, trainControl.repeats = glb_rcv_n_repeats,
        trainControl.classProbs = glb_is_classification,
        trainControl.summaryFunction = glbMdlMetricSummaryFn,
        trainControl.allowParallel = glbMdlAllowParallel,                                
        train.metric = glbMdlMetricSummary, 
        train.maximize = glbMdlMetricMaximize,    
        train.method = "glmnet")),
        indepVar = indepVars,
        rsp_var = glb_rsp_var, 
        fit_df = glbObsFit, OOB_df = glbObsOOB)

    indepVars <- c(max_cor_y_x_vars)
    for (txtFeat in names(glbFeatsText))
        indepVars <- union(indepVars, 
            grep(paste(str_to_upper(substr(txtFeat, 1, 1)), "\\.T\\.", sep = ""),
                        names(glbObsAll), perl = TRUE, value = TRUE))
    indepVars <- myadjustInteractionFeats(glb_feats_df, indepVars)
    ret_lst <- myfit_mdl(mdl_specs_lst = myinit_mdl_specs_lst(mdl_specs_lst = list(
        id.prefix = "Max.cor.Y.Text.onlyT", 
        type = glb_model_type, 
        tune.df = glbMdlTuneParams,        
        trainControl.method = "repeatedcv",
        trainControl.number = glb_rcv_n_folds, trainControl.repeats = glb_rcv_n_repeats,
        trainControl.classProbs = glb_is_classification,
        trainControl.summaryFunction = glbMdlMetricSummaryFn,
        train.metric = glbMdlMetricSummary, 
        train.maximize = glbMdlMetricMaximize,    
        train.method = "glmnet")),
        indepVar = indepVars,
        rsp_var = glb_rsp_var, 
        fit_df = glbObsFit, OOB_df = glbObsOOB)

    indepVars <- c(max_cor_y_x_vars)
    for (txtFeat in names(glbFeatsText))
        indepVars <- union(indepVars, 
            grep(paste(str_to_upper(substr(txtFeat, 1, 1)), "\\.P\\.", sep = ""),
                        names(glbObsAll), perl = TRUE, value = TRUE))
    indepVars <- myadjustInteractionFeats(glb_feats_df, indepVars)
    ret_lst <- myfit_mdl(mdl_specs_lst = myinit_mdl_specs_lst(mdl_specs_lst = list(
        id.prefix = "Max.cor.Y.Text.onlyP", 
        type = glb_model_type, 
        tune.df = glbMdlTuneParams,        
        trainControl.method = "repeatedcv",
        trainControl.number = glb_rcv_n_folds, trainControl.repeats = glb_rcv_n_repeats,
        trainControl.classProbs = glb_is_classification,
        trainControl.summaryFunction = glbMdlMetricSummaryFn,
        trainControl.allowParallel = glbMdlAllowParallel,        
        train.metric = glbMdlMetricSummary, 
        train.maximize = glbMdlMetricMaximize,    
        train.method = "glmnet")),
        indepVar = indepVars,
        rsp_var = glb_rsp_var, 
        fit_df = glbObsFit, OOB_df = glbObsOOB)
}

# Interactions.High.cor.Y
if (length(int_feats <- setdiff(setdiff(unique(glb_feats_df$cor.high.X), NA), 
                                subset(glb_feats_df, nzv)$id)) > 0) {
    fit.models_0_chunk_df <- myadd_chunk(fit.models_0_chunk_df, 
                    paste0("fit.models_0_", "Interact.High.cor.Y"), major.inc = FALSE,
                                    label.minor = "glmnet")

    ret_lst <- myfit_mdl(mdl_specs_lst=myinit_mdl_specs_lst(mdl_specs_lst=list(
        id.prefix="Interact.High.cor.Y", 
        type=glb_model_type, trainControl.method="repeatedcv",
        trainControl.number=glb_rcv_n_folds, trainControl.repeats=glb_rcv_n_repeats,
        trainControl.classProbs = glb_is_classification,
        trainControl.summaryFunction = glbMdlMetricSummaryFn,
        trainControl.allowParallel = glbMdlAllowParallel,
        train.metric = glbMdlMetricSummary, 
        train.maximize = glbMdlMetricMaximize,    
        train.method="glmnet")),
        indepVar=c(max_cor_y_x_vars, paste(max_cor_y_x_vars[1], int_feats, sep=":")),
        rsp_var=glb_rsp_var, 
        fit_df=glbObsFit, OOB_df=glbObsOOB)
}    

# Low.cor.X
fit.models_0_chunk_df <- myadd_chunk(fit.models_0_chunk_df, 
                        paste0("fit.models_0_", "Low.cor.X"), major.inc = FALSE,
                                     label.minor = "glmnet")
indepVar <- mygetIndepVar(glb_feats_df)
indepVar <- setdiff(indepVar, unique(glb_feats_df$cor.high.X))
ret_lst <- myfit_mdl(mdl_specs_lst = myinit_mdl_specs_lst(mdl_specs_lst = list(
            id.prefix = "Low.cor.X", 
            type = glb_model_type, 
            tune.df = glbMdlTuneParams,        
            trainControl.method = "repeatedcv",
            trainControl.number = glb_rcv_n_folds, trainControl.repeats = glb_rcv_n_repeats,
            trainControl.classProbs = glb_is_classification,
            trainControl.summaryFunction = glbMdlMetricSummaryFn,
            trainControl.allowParallel = glbMdlAllowParallel,
            train.metric = glbMdlMetricSummary, 
            train.maximize = glbMdlMetricMaximize,    
            train.method = "glmnet")),
        indepVar = indepVar, rsp_var = glb_rsp_var, 
        fit_df = glbObsFit, OOB_df = glbObsOOB)

fit.models_0_chunk_df <- 
    myadd_chunk(fit.models_0_chunk_df, "fit.models_0_end", major.inc = FALSE,
                label.minor = "teardown")

rm(ret_lst)

glb_chunks_df <- myadd_chunk(glb_chunks_df, "fit.models", major.inc = FALSE)

```{r fit.models_1, cache=FALSE, fig.height=10, fig.width=15, eval=myevlChunk(glbChunks, glbOut$pfx)}

```{r fit.models_2, cache=FALSE, fig.height=10, fig.width=15, eval=myevlChunk(glbChunks, glbOut$pfx)}

# if (sum(is.na(glbObsAll$D.P.http)) > 0)
#         stop("fit.models_3: Why is this happening ?")

#stop(here"); glb2Sav()
sync_glb_obs_df <- function() {
    # Merge or cbind ?
    for (col in setdiff(names(glbObsFit), names(glbObsTrn)))
        glbObsTrn[glbObsTrn$.lcn == "Fit", col] <<- glbObsFit[, col]
    for (col in setdiff(names(glbObsFit), names(glbObsAll)))
        glbObsAll[glbObsAll$.lcn == "Fit", col] <<- glbObsFit[, col]
    if (all(is.na(glbObsNew[, glb_rsp_var])))
        for (col in setdiff(names(glbObsOOB), names(glbObsTrn)))
            glbObsTrn[glbObsTrn$.lcn == "OOB", col] <<- glbObsOOB[, col]
    for (col in setdiff(names(glbObsOOB), names(glbObsAll)))
        glbObsAll[glbObsAll$.lcn == "OOB", col] <<- glbObsOOB[, col]
}
sync_glb_obs_df()
    
print(setdiff(names(glbObsNew), names(glbObsAll)))

replay.petrisim(pn = glb_analytics_pn, 
    replay.trans = (glb_analytics_avl_objs <- c(glb_analytics_avl_objs, 
        "model.selected")), flip_coord = TRUE)
glb_chunks_df <- myadd_chunk(glb_chunks_df, "fit.data.training", major.inc = TRUE)

Step 5.0: cluster data

```{r fit.data.training_0, cache=FALSE, eval=myevlChunk(glbChunks, glbOut$pfx)}

#stop(here"); glb2Sav()
if (glb_is_classification && glb_is_binomial) 
    prob_threshold <- glb_models_df[glb_models_df$id == glbMdlSelId,
                                        "opt.prob.threshold.OOB"] else 
    prob_threshold <- NULL

if (grepl("Ensemble", glbMdlFinId)) {
    # Get predictions for each model in ensemble; Outliers that have been moved to OOB might not have been predicted yet
    mdlEnsembleComps <- unlist(str_split(subset(glb_models_df, 
                                                id == glbMdlFinId)$feats, ","))
    if (glb_is_classification)
    #     mdlEnsembleComps <- gsub("\\.prob$", "", mdlEnsembleComps)
    # mdlEnsembleComps <- gsub(paste0("^", 
    #                     gsub(".", "\\.", mygetPredictIds(glb_rsp_var)$value, fixed = TRUE)),
    #                          "", mdlEnsembleComps)
        mdlEnsembleComps <- glb_models_df$id[sapply(glb_models_df$id, function(thsMdlId)
                        mygetPredictIds(glb_rsp_var, thsMdlId)$prob  %in% mdlEnsembleComps)] else
        mdlEnsembleComps <- glb_models_df$id[sapply(glb_models_df$id, function(thsMdlId)
                        mygetPredictIds(glb_rsp_var, thsMdlId)$value  %in% mdlEnsembleComps)]
                        
    for (mdl_id in mdlEnsembleComps) {
        glbObsTrn <- glb_get_predictions(df = glbObsTrn, mdl_id = mdl_id, 
                                            rsp_var = glb_rsp_var,
                                            prob_threshold_def = prob_threshold)
        glbObsNew <- glb_get_predictions(df = glbObsNew, mdl_id = mdl_id, 
                                            rsp_var = glb_rsp_var,
                                            prob_threshold_def = prob_threshold)
        # glb_fin_mdl uses the same coefficients as glb_sel_mdl, 
        #   so copy the "Final" columns into "non-Final" columns
        glbObsTrn[, gsub("Final.", "", unlist(mygetPredictIds(glb_rsp_var, mdl_id)))] <-
            glbObsTrn[, unlist(mygetPredictIds(glb_rsp_var, mdl_id))]
        glbObsNew[, gsub("Final.", "", unlist(mygetPredictIds(glb_rsp_var, mdl_id)))] <-
            glbObsNew[, unlist(mygetPredictIds(glb_rsp_var, mdl_id))]
    }    
}
glbObsTrn <- glb_get_predictions(df = glbObsTrn, mdl_id = glbMdlFinId, 
                                     rsp_var = glb_rsp_var,
                                    prob_threshold_def = prob_threshold)

glb_featsimp_df <- myget_feats_importance(mdl=glb_fin_mdl,
                                          featsimp_df=glb_featsimp_df)
#glb_featsimp_df[, paste0(glbMdlFinId, ".imp")] <- glb_featsimp_df$imp
print(glb_featsimp_df)
if (glb_is_classification && glb_is_binomial)
    glb_analytics_diag_plots(obs_df=glbObsTrn, mdl_id=glbMdlFinId, 
            prob_threshold=glb_models_df[glb_models_df$id == glbMdlSelId, 
                                         "opt.prob.threshold.OOB"]) else
    glb_analytics_diag_plots(obs_df=glbObsTrn, mdl_id=glbMdlFinId)                  

dsp_feats_vctr <- c(NULL)
for(var in grep(".imp", names(glb_feats_df), fixed=TRUE, value=TRUE))
    dsp_feats_vctr <- union(dsp_feats_vctr, 
                            glb_feats_df[!is.na(glb_feats_df[, var]), "id"])

# print(glbObsTrn[glbObsTrn$UniqueID %in% FN_OOB_ids, 
#                     grep(glb_rsp_var, names(glbObsTrn), value=TRUE)])

print(setdiff(names(glbObsTrn), names(glbObsAll)))
for (col in setdiff(names(glbObsTrn), names(glbObsAll)))
    # Merge or cbind ?
    glbObsAll[glbObsAll$.src == "Train", col] <- glbObsTrn[, col]

print(setdiff(names(glbObsFit), names(glbObsAll)))
print(setdiff(names(glbObsOOB), names(glbObsAll)))
for (col in setdiff(names(glbObsOOB), names(glbObsAll)))
    # Merge or cbind ?
    glbObsAll[glbObsAll$.lcn == "OOB", col] <- glbObsOOB[, col]
    
print(setdiff(names(glbObsNew), names(glbObsAll)))

#glb2Sav(); all.equal(savObsAll, glbObsAll); all.equal(sav_models_lst, glb_models_lst)
#load(file = paste0(glbOut$pfx, "dsk_knitr.RData"))
#cmpCols <- names(glbObsAll)[!grepl("\\.Final\\.", names(glbObsAll))]; all.equal(savObsAll[, cmpCols], glbObsAll[, cmpCols]); all.equal(savObsAll[, "H.P.http"], glbObsAll[, "H.P.http"]); 

replay.petrisim(pn = glb_analytics_pn, 
    replay.trans = (glb_analytics_avl_objs <- c(glb_analytics_avl_objs, 
        "data.training.all.prediction","model.final")), flip_coord = TRUE)
glb_chunks_df <- myadd_chunk(glb_chunks_df, "predict.data.new", major.inc = TRUE)

Step 5.0: cluster data

Null Hypothesis (\(\sf{H_{0}}\)): mpg is not impacted by am_fctr.
The variance by am_fctr appears to be independent. #{r q1, cache=FALSE} # print(t.test(subset(cars_df, am_fctr == "automatic")$mpg, # subset(cars_df, am_fctr == "manual")$mpg, # var.equal=FALSE)$conf) # We reject the null hypothesis i.e. we have evidence to conclude that am_fctr impacts mpg (95% confidence). Manual transmission is better for miles per gallon versus automatic transmission.

##                        label step_major step_minor label_minor     bgn
## 2               inspect.data          2          0           0  11.999
## 3                 scrub.data          2          1           1 145.666
## 1                import.data          1          0           0   5.768
## 11      extract.features.end          3          6           6 180.944
## 12       manage.missing.data          4          0           0 181.818
## 10   extract.features.string          3          5           5 180.875
## 9      extract.features.text          3          4           4 180.818
## 7     extract.features.image          3          2           2 180.732
## 4             transform.data          2          2           2 180.630
## 6  extract.features.datetime          3          1           1 180.692
## 8     extract.features.price          3          3           3 180.783
## 5           extract.features          3          0           0 180.671
##        end elapsed duration
## 2  145.665 133.666  133.666
## 3  180.630  34.964   34.964
## 1   11.999   6.231    6.231
## 11 181.818   0.874    0.874
## 12 182.334   0.516    0.516
## 10 180.943   0.069    0.068
## 9  180.875   0.057    0.057
## 7  180.782   0.051    0.050
## 4  180.671   0.041    0.041
## 6  180.731   0.039    0.039
## 8  180.817   0.034    0.034
## 5  180.691   0.020    0.020
## [1] "Total Elapsed Time: 182.334 secs"